EmbeddingDropouts
EmbeddingLayerNorms
TransformerEmbeddings
TransformerDropouts
TransformerLayerNorms
DecoderLayer
EncoderLayer
QkvMode
QkvSplit
AttentionHeads
AttentionMask
KeyValueCache
AttentionScorer
AttentionLinearBiases
ScaledDotProductAttention
SelfAttention
SinusoidalPositionalEmbedding
RotaryEmbeddings
QueryKeyRotaryEmbeddings
PointwiseFeedForward
Activation
GELUFast
GELUNew
RMSNorm
ModelOutput
ModelOutputWithCache
CausalLMOutputWithCache
RotaryEmbeddingConfig
TransformerAttentionLayerConfig
TransformerEmbeddingLayerConfig
TransformerFeedForwardLayerConfig
TransformerLayerConfig
TransformerConfig
EncoderModule
TransformerEncoder
ALBERTEncoder
BERTEncoder
CamemBERTEncoder
RoBERTaEncoder
XLMREncoder
AutoEncoder
DecoderModule
TransformerDecoder
FalconDecoder
GPTNeoXDecoder
LlamaDecoder
MPTDecoder
AutoDecoder
CausalLMModule
TransformerCausalLM
FalconCausalLM
GPTNeoXCausalLM
LlamaCausalLM
MPTCausalLM
AutoCausalLM
Generator
StringGenerator
GeneratorWrapper
DefaultGenerator
DollyV2Generator
FalconGenerator
LlamaGenerator
MPTGenerator
AutoGenerator
GeneratorConfig
GreedyGeneratorConfig
SampleGeneratorConfig
StopCondition
CompoundStopCondition
EndOfSequenceCondition
MaxGeneratedPiecesCondition
LogitsTransform
CompoundLogitsTransform
TopKTransform
TopPTransform
TemperatureTransform
VocabMaskTransform
Repository
RepositoryFile
TransactionContext
FsspecRepository
HfHubRepository
FsspecFile
LocalFile
HfHubFile
InputChunks
SpecialPieceChunk
TextChunk
PiecesWithIds
AutoTokenizer
TokenizerBase
Quantizable
Quantizable.modules_to_not_quantize()
Dtype4Bit
BitsAndBytesConfig
enable_torch_sdp()
use_nvtx_ranges_for_forward_pass()
Default
FutureMandatory