Byte Goose AI Docs
DeepSeek ENGRAM vs SSM Mamba
Initializing search
Byte Goose AI Docs
Home
Podcasts
Podcasts
DeepSeek ENGRAM vs SSM Mamba
Text-generation Models
Text-generation Models
Apertus
BLOOM
Chinese-Vicuna
Claude Opus 4.5
Code Llama
DASD-4B-Thinking
DeepSeek-V3
DeepSeek-V3.2
GLM-4.5
GPT-4
GPT-5
Gemini 3 Pro
Gemma 3
Grok 4
Kimi K2
Llama 4
MiniMax-M1
OpenAI o1
Qwen3
WaveCoder
phi-3-mini
Image-generation Models
Image-generation Models
DALL·E 3
FLUX.1 Kontext
Imagen
InfiniteYou
Nano Banana Pro
PIXARTδ
Playground v3
RealGen
SDXL
Seedream 3.0
Stable Diffusion 3.5 Large
Z-Image
Fine tuning
Fine tuning
RL Fine-tuning
RL Fine-tuning
AlignProp
Binary Classifier Optimization
Contrastive Preference Optimization
Direct Preference Optimization
Explicit Policy Optimization
Exploratory Preference Optimization
GRPO
GRPO-RM
Generalized Contrastive Learning
Gradient Knowledge Distillation (GKD)
Guided Pivotal Optimization
Implicit Language Q-learning
Iterative Foundation Model Fine-Tuning on Multiple Rewards
KTO
Location Preference Optimization
MTR-DuplexBench
Multi-Objective Hyperparameter Optimization
Nabla-R2D3
Nash learning from human feedback
Natural Policy Gradient
OLLIE
ORPO
Oracle-RLAIF
Orthogonal Finetuning for Direct Preference Optimization
PEARL
Proximal Supervised Fine-Tuning
Q-SFT
R-PRM
RLHF Fine-Tuning
RLOO
Saliency-Aware Quantized Imitation Learning
Self-NPO
Self-Rewarding PPO
T-REX
TD3-C
Vision-Driven Prompt Optimization
PEFT
PEFT
DeepSeek ENGRAM vs SSM Mamba