Matt commited on
Commit
f18f828
·
1 Parent(s): bb44b80

Update model and modeling file

Browse files
Files changed (2) hide show
  1. model.safetensors +3 -0
  2. modeling_florence2.py +3 -21
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b4e610c952eef90a836c56cda0f398a672a3a6ca7b4d96b0e09a86dee42e2c3
3
+ size 1540980506
modeling_florence2.py CHANGED
@@ -26,9 +26,10 @@ import torch.utils.checkpoint as checkpoint
26
  from torch.nn import CrossEntropyLoss
27
  from collections import OrderedDict
28
  from einops import rearrange
29
- from timm.models.layers import DropPath, trunc_normal_
30
 
31
  from transformers.modeling_utils import PreTrainedModel
 
32
  from transformers.utils import (
33
  ModelOutput,
34
  add_start_docstrings,
@@ -609,29 +610,10 @@ class DaViT(nn.Module):
609
  self.avgpool = nn.AdaptiveAvgPool1d(1)
610
  self.head = nn.Linear(self.embed_dims[-1], num_classes) if num_classes > 0 else nn.Identity()
611
 
612
- self.apply(self._init_weights)
613
-
614
  @property
615
  def dim_out(self):
616
  return self.embed_dims[-1]
617
 
618
- def _init_weights(self, m):
619
- if isinstance(m, nn.Linear):
620
- trunc_normal_(m.weight, std=0.02)
621
- if m.bias is not None:
622
- nn.init.constant_(m.bias, 0)
623
- elif isinstance(m, nn.Conv2d):
624
- nn.init.normal_(m.weight, std=0.02)
625
- for name, _ in m.named_parameters():
626
- if name in ['bias']:
627
- nn.init.constant_(m.bias, 0)
628
- elif isinstance(m, nn.LayerNorm):
629
- nn.init.constant_(m.weight, 1.0)
630
- nn.init.constant_(m.bias, 0)
631
- elif isinstance(m, nn.BatchNorm2d):
632
- nn.init.constant_(m.weight, 1.0)
633
- nn.init.constant_(m.bias, 0)
634
-
635
  def forward_features_unpool(self, x):
636
  """
637
  forward until avg pooling
@@ -2059,7 +2041,7 @@ class Florence2LanguageModel(Florence2LanguagePreTrainedModel):
2059
  )
2060
 
2061
 
2062
- class Florence2LanguageForConditionalGeneration(Florence2LanguagePreTrainedModel):
2063
  base_model_prefix = "model"
2064
  _tied_weights_keys = ["encoder.embed_tokens.weight", "decoder.embed_tokens.weight", "lm_head.weight"]
2065
  _keys_to_ignore_on_load_missing = ["final_logits_bias"]
 
26
  from torch.nn import CrossEntropyLoss
27
  from collections import OrderedDict
28
  from einops import rearrange
29
+ from timm.layers import DropPath, trunc_normal_
30
 
31
  from transformers.modeling_utils import PreTrainedModel
32
+ from transformers.generation.utils import GenerationMixin
33
  from transformers.utils import (
34
  ModelOutput,
35
  add_start_docstrings,
 
610
  self.avgpool = nn.AdaptiveAvgPool1d(1)
611
  self.head = nn.Linear(self.embed_dims[-1], num_classes) if num_classes > 0 else nn.Identity()
612
 
 
 
613
  @property
614
  def dim_out(self):
615
  return self.embed_dims[-1]
616
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
617
  def forward_features_unpool(self, x):
618
  """
619
  forward until avg pooling
 
2041
  )
2042
 
2043
 
2044
+ class Florence2LanguageForConditionalGeneration(Florence2LanguagePreTrainedModel, GenerationMixin):
2045
  base_model_prefix = "model"
2046
  _tied_weights_keys = ["encoder.embed_tokens.weight", "decoder.embed_tokens.weight", "lm_head.weight"]
2047
  _keys_to_ignore_on_load_missing = ["final_logits_bias"]