Upload SAINEMO-reMIX GPTQ 8-bit model files
Browse files- .gitattributes +1 -0
- config.json +182 -0
- generation_config.json +7 -0
- model-00001-of-00003.safetensors +3 -0
- model-00002-of-00003.safetensors +3 -0
- model-00003-of-00003.safetensors +3 -0
- model.safetensors.index.json +1210 -0
- special_tokens_map.json +30 -0
- tokenizer.json +3 -0
- tokenizer_config.json +0 -0
.gitattributes
CHANGED
|
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
config.json
ADDED
|
@@ -0,0 +1,182 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "Moraliane/SAINEMO-reMIX",
|
| 3 |
+
"architectures": [
|
| 4 |
+
"MistralForCausalLM"
|
| 5 |
+
],
|
| 6 |
+
"attention_dropout": 0.0,
|
| 7 |
+
"bos_token_id": 1,
|
| 8 |
+
"eos_token_id": 2,
|
| 9 |
+
"head_dim": 128,
|
| 10 |
+
"hidden_act": "silu",
|
| 11 |
+
"hidden_size": 5120,
|
| 12 |
+
"initializer_range": 0.02,
|
| 13 |
+
"intermediate_size": 14336,
|
| 14 |
+
"max_position_embeddings": 1024000,
|
| 15 |
+
"model_type": "mistral",
|
| 16 |
+
"num_attention_heads": 32,
|
| 17 |
+
"num_hidden_layers": 40,
|
| 18 |
+
"num_key_value_heads": 8,
|
| 19 |
+
"pad_token_id": 10,
|
| 20 |
+
"quantization_config": {
|
| 21 |
+
"batch_size": 1,
|
| 22 |
+
"bits": 8,
|
| 23 |
+
"block_name_to_quantize": null,
|
| 24 |
+
"cache_block_outputs": true,
|
| 25 |
+
"damp_percent": 0.01,
|
| 26 |
+
"dataset": [
|
| 27 |
+
" = = = Religious works = = = \n",
|
| 28 |
+
" = = Milestones = = \n",
|
| 29 |
+
" Note : GP \n",
|
| 30 |
+
" = = Personal life and other work = = \n",
|
| 31 |
+
" = = Early life and modeling career = = \n",
|
| 32 |
+
" Like other stingrays , the plain maskray is viviparous with the developing embryos sustained to term by histotroph ( \" uterine milk \" ) produced by the mother . Mature females have a single functional ovary and uterus , on the left . Litter size is one or two ; the newborns measure 12 \u2013 14 cm ( 4 @.@ 7 \u2013 5 @.@ 5 in ) across . Males and females reach sexual maturity at disc widths of 20 \u2013 21 cm ( 7 @.@ 9 \u2013 8 @.@ 3 in ) and 18 \u2013 19 cm ( 7 @.@ 1 \u2013 7 @.@ 5 in ) respectively . The maximum lifespan is estimated to be 9 years for males and 13 years for females . \n",
|
| 33 |
+
" The Children \u2019 s Book of Hymns ; Blackie , 1929 ; rep . 1933 \n",
|
| 34 |
+
" M1822 .69 cal ( percussion @-@ converted ) 53 \n",
|
| 35 |
+
" = = In the media = = \n",
|
| 36 |
+
" = = = Book covers = = = \n",
|
| 37 |
+
" The pectoral fin disc of the plain maskray is thin and diamond @-@ shaped with narrowly rounded outer corners , measuring 1 @.@ 1 \u2013 1 @.@ 3 times longer than wide . The leading margins of the disc are gently concave and converge at a broad angle to the pointed tip of the snout . The small eyes are placed close together , and behind them are the spiracles . The nostrils are elongated and have a skirt @-@ shaped flap of skin between them . The small mouth bears prominent furrows at the corners and contains two slender papillae on the floor . Small papillae are also found around the outside of the mouth . There are five pairs of gill slits . The pelvic fins are fairly large and pointed . \n",
|
| 38 |
+
" The Flower Fairy Picture Book ; Blackie , 1955 \n",
|
| 39 |
+
" The accompanying music video begins with a shot of an empty street , followed by clips of disadvantaged and poorer members of society going about their daily activities . Two men play dominoes on a wooden crate outside a building , a gang make fun of an elderly man hanging newspapers outside his store and an obese woman walks down the street . Clips of Carey leaning against a wall and sitting on some steps looking on at what is happening are shown . As the first chorus begins , everyone starts to dance joyfully in the street and help those in need . A gospel choir comes out of one of the buildings as the street becomes more crowded with people of all ages and backgrounds rejoicing and getting along with each other . One of the shops in the background has a neon light outside the entrance which says \" Jesus Saves \" . \n",
|
| 40 |
+
" Simon the Swan ; Blackie , 1988 \n",
|
| 41 |
+
" Flower Fairies of the Seasons ; Bedrick / Blackie , 1988 \n",
|
| 42 |
+
" One of the most popular actresses in India , she was the recipient of the IIFA Award for Star Debut of the Year \u2013 Female in 2010 . Alongside her screen acting career , Fernandez has participated in stage shows , and is active in humanitarian work . \n",
|
| 43 |
+
" As the Nameless officially do not exist , the upper echelons of the Gallian Army exploit the concept of plausible deniability in order to send them on missions that would otherwise make Gallia lose face in the war . While at times this works to their advantage , such as a successful incursion into Imperial territory , other orders cause certain members of the 422nd great distress . One such member , Gusurg , becomes so enraged that he abandons his post and defects into the ranks of Calamity Raven , attached to the ideal of Darcsen independence proposed by their leader , Dahau . At the same time , elements within Gallian Army Command move to erase the Nameless in order to protect their own interests . Hounded by both allies and enemies , and combined with the presence of a traitor within their ranks , the 422nd desperately move to keep themselves alive while at the same time fight to help the Gallian war effort . This continues until the Nameless 's commanding officer , Ramsey Crowe , who had been kept under house arrest , is escorted to the capital city of Randgriz in order to present evidence exonerating the weary soldiers and expose the real traitor , the Gallian General that had accused Kurt of Treason . \n",
|
| 44 |
+
" Repaired : \n",
|
| 45 |
+
" A month later , on May 14 , the Blue Jackets announced that Richards would remain as head coach and signed him to a two @-@ year contract . During the press conference , Howson noted , \" Our team continuously improved under Todd and he has earned the opportunity to build upon the work he started . \" Columbus posted an 18 \u2013 21 \u2013 2 record under Richards , including winning seven of their final 11 games . \n",
|
| 46 |
+
" \" There 's Got to Be a Way \" ( 7 \" remix ) \n",
|
| 47 |
+
" Received & Issued : \n",
|
| 48 |
+
" In the late 1920s , Barker began to doubt she was doing enough for the church and considered focusing solely on sacred works . Family and friends recommended she continue secular and sacred works , which she did . \n",
|
| 49 |
+
" The Gambia also has an under @-@ 19 team that was to play in the African Women 's U @-@ 19 Championship in 2002 . The Gambia 's first match was against Morocco , but the team withdrew from the competition . \n",
|
| 50 |
+
" Cullen was a standout with BU ; he was named the East Coast Athletic Conference Rookie of the Year in 1983 \u2013 84 after leading his team in scoring with 56 points . The National Hockey League passed him over , however , as he went unclaimed in the 1984 NHL Entry Draft . He was named to the Hockey East All @-@ Star Teams in 1985 , 1986 and 1987 , and a National Collegiate Athletic Association East Second Team All @-@ American in 1986 . He graduated as BU 's all @-@ time scoring leader with 241 points , and was named to BU 's Hockey East 25th anniversary team in 2009 . \n",
|
| 51 |
+
" A Child \u2019 s Garden of Verses ( Robert Louis Stevenson ) ; Blackie , 1944 \n",
|
| 52 |
+
" Picturesque Children of the Allies ; J. Salmon , 1916 \n",
|
| 53 |
+
" = = = Pre @-@ season = = = \n",
|
| 54 |
+
" Barker was equally proficient in watercolour , pen and ink , oils , and pastels . Kate Greenaway and the Pre @-@ Raphaelites were the principal influences on her work . She claimed to paint instinctively and rejected artistic theories . Barker died in 1973 . Though she published Flower Fairy books with spring , summer , and autumn themes , it wasn 't until 1985 that a winter collection was assembled from her remaining work and published posthumously . \n",
|
| 55 |
+
" The first scientific description of the plain maskray was authored by Commonwealth Scientific and Industrial Research Organisation ( CSIRO ) researcher Peter Last in a 1987 issue of Memoirs of the National Museum of Victoria . The specific name annotatus comes from the Latin an ( \" not \" ) and notatus ( \" marked \" ) , and refers to the ray 's coloration . The holotype is a male 21 @.@ 2 cm ( 8 @.@ 3 in ) across , caught off Western Australia ; several paratypes were also designated . Last tentatively placed the species in the genus Dasyatis , noting that it belonged to the \" maskray \" species group that also included the bluespotted stingray ( then Dasyatis kuhlii ) . In 2008 , Last and William White elevated the kuhlii group to the rank of full genus as Neotrygon , on the basis of morphological and molecular phylogenetic evidence . \n",
|
| 56 |
+
" \" There 's Got to Be a Way \" ( Original album version ) \u2013 4 : 52 \n",
|
| 57 |
+
" 14 @,@ 000 buck & ball cartridges - flint \n",
|
| 58 |
+
" = = Filmography = = \n",
|
| 59 |
+
" = = Schedule and results = = \n",
|
| 60 |
+
" = = Standings = = \n",
|
| 61 |
+
" Note : Pos \n",
|
| 62 |
+
" In 1911 , Raphael Tuck & Sons bought four of Barker 's \" little drawings \" for half a sovereign , and published them as postcards . In October 1911 , she won second prize in the Croydon Art Society 's poster competition , and shortly afterward was elected the youngest member of the Society . The art critic for the Croydon Advertiser remarked , \" Her drawings show a remarkable freedom of spirit . She has distinct promise . \" \n",
|
| 63 |
+
" = = Fourth stage = = \n",
|
| 64 |
+
" A Little Book of Prayers and Hymns ; Frederick Warne , 1994 \n",
|
| 65 |
+
" Beautiful Bible Pictures ; Blackie , 1932 \n",
|
| 66 |
+
" In 1940 , the Barker 's live @-@ in maid retired , and Dorothy Barker closed her school at the back of the house in The Waldrons . She continued to supervise the household , and to give both her mother and sister the care they needed . Dorothy and her sister collaborated upon only two books : Our Darling 's First Book and the Christian @-@ themed , He Leadeth Me . In 1954 Dorothy Barker died of a heart attack . Barker was unable to pursue her art to any significant extent following her sister 's death , as all the care of her aged mother devolved upon her , but she did manage to begin planning a stained glass window design in her sister 's memory for St. Edmund 's , Pitlake . \n",
|
| 67 |
+
" Flower Fairies of the Garden ; Blackie , 1944 \n",
|
| 68 |
+
" Seaside Holiday ; J. Salmon , 1918 , 1921 \n",
|
| 69 |
+
" = = Major intersections = = \n",
|
| 70 |
+
" = = Distribution and habitat = = \n",
|
| 71 |
+
" Old Rhymes for All Times ; Blackie , 1928 \n",
|
| 72 |
+
" The game began development in 2010 , carrying over a large portion of the work done on Valkyria Chronicles II . While it retained the standard features of the series , it also underwent multiple adjustments , such as making the game more forgiving for series newcomers . Character designer Raita Honjou and composer Hitoshi Sakimoto both returned from previous entries , along with Valkyria Chronicles II director Takeshi Ozawa . A large team of writers handled the script . The game 's opening theme was sung by May 'n . \n",
|
| 73 |
+
" = = Off @-@ season = = \n",
|
| 74 |
+
" 2 @,@ 236 shotguns and rifles ( repaired mostly for troops in service ) \n",
|
| 75 |
+
" How to find Flower Fairies ; Frederick Warne , 2007 \n",
|
| 76 |
+
" A New Epiphany ; Society for the Preservation of Christian Knowledge , 1919 \n",
|
| 77 |
+
" Valkyria Chronicles 3 was adapted into a two @-@ episode original video animation series in the same year of its release . Titled Senj\u014d no Valkyria 3 : Taga Tame no J\u016bs\u014d ( \u6226\u5834\u306e\u30f4\u30a1\u30eb\u30ad\u30e5\u30ea\u30a2\uff13 \u8ab0\u304c\u305f\u3081\u306e\u9283\u7621 , lit . Valkyria of the Battlefield 3 : The Wound Taken for Someone 's Sake ) , it was originally released through PlayStation Network and Qriocity between April and May 2011 . The initially @-@ planned release and availability period needed to be extended due to a stoppage to PSN during the early summer of that year . It later released for DVD on June 29 and August 31 , 2011 , with separate \" Black \" and \" Blue \" editions being available for purchase . The anime is set during the latter half of Valkyria Chronicles III , detailing a mission by the Nameless against their Imperial rivals Calamity Raven . The anime was first announced in November 2010 . It was developed by A @-@ 1 Pictures , produced by Shinji Motoyama , directed by Nobuhiro Kond\u014d , and written by Hiroshi \u014cnogi . Sakimoto 's music for the game was used in the anime . \n",
|
| 78 |
+
" = = = Release = = = \n",
|
| 79 |
+
" A Little Book of Old Rhymes ; Blackie , 1936 \n",
|
| 80 |
+
" In 1942 , the Tower Building was renovated due to the efforts of the \u00c6sthetic Club , Little Rock philanthropist Frederick W. Allsop , and the Works Progress Administration . It became the new home of The Arkansas Museum of Natural History and Antiquities , which had been located in Little Rock City Hall . The museum remained in the tower building for approximately fifty @-@ five years . The area surrounding the Tower Building had been known as Arsenal Park when the first decommissioned and then later renamed City Park . Due to the efforts of Bernie Babcock , however , the city finally named it MacArthur Park in 1942 in honor of Douglas MacArthur . \n",
|
| 81 |
+
" \" There 's Got to Be a Way \" is an R & B @-@ pop music song with elements of gospel . The theme of social activism can be heard in the lyrics \" There \u2019 s got to be a way / to connect this world today . \" The song begins with Carey publicly denouncing the existence of poverty and racism in the world , and she uses the bridge to shift the lyrics towards an uplifting and aspirational tone . Carey suggests we should be more tolerant of each other and not resort so readily to war in the lyrics \" Couldn 't we accept each other / Can 't we make ourselves aware . \" \n",
|
| 82 |
+
" = = Music video = = \n",
|
| 83 |
+
" Divisions : CE \u2013 Central , NW \u2013 Northwest , PA \u2013 Pacific \n",
|
| 84 |
+
" = = Player statistics = = \n",
|
| 85 |
+
" Following her father \u2019 s death in June 1912 , the seventeen @-@ year @-@ old Barker submitted art and poetry to My Magazine , Child \u2019 s Own , Leading Strings , and Raphael Tuck annuals in an effort to support both her mother and sister . Her sister Dorothy taught kindergarten in two private schools before opening a kindergarten at home . She brought in some money for the family 's support while supervising the household . \n",
|
| 86 |
+
" St. Cecily 's Garden ; 1920 \n",
|
| 87 |
+
" Most of the equipment , arms , and machinery at the Little Rock Arsenal was removed to east of the Mississippi River by order of Maj. Gen. Earl Van Dorn in April and May 1862 , and accountability for it is lost at that point . By all appearances , the equipment was sent down the river to Napoleon , Arkansas , and from there to Jackson Mississippi , where it was probably destroyed during the Vicksburg campaign in the early summer of 1863 . \n",
|
| 88 |
+
" M1822 .69 cal ( flintlock ) 5 @,@ 625 \n",
|
| 89 |
+
" In 1864 , after Little Rock fell to the Union Army and the arsenal had been recaptured , General Fredrick Steele marched 8 @,@ 500 troops from the arsenal beginning the Camden Expedition . \n",
|
| 90 |
+
" In the early evening of 15 June , Webb began his next attempt by passing over the lower obstructions in the Wilmington River and spent the rest of the night coaling . He moved forward the next evening to a concealed position within easy reach of the monitors for an attack early the following morning . Webb planned to sink one of the monitors with his spar torpedo and then deal with the other one with his guns . The gunboat Isondiga and the tugboat Resolute were to accompany him to tow one or both of the monitors back to Savannah . \n",
|
| 91 |
+
" Two manga adaptations were produced , following each of the game 's main female protagonists Imca and Riela . They were Senj\u014d no Valkyria 3 : Namo naki Chikai no Hana ( \u6226\u5834\u306e\u30f4\u30a1\u30eb\u30ad\u30e5\u30ea\u30a23 \u540d\u3082\u306a\u304d\u8a93\u3044\u306e\u82b1 , lit . Valkyria of the Battlefield 3 : The Flower of the Nameless Oath ) , illustrated by Naoyuki Fujisawa and eventually released in two volumes after being serialized in Dengeki Maoh between 2011 and 2012 ; and Senj\u014d no Valkyria 3 : -Akaki Unmei no Ikusa Otome- ( \u6226\u5834\u306e\u30f4\u30a1\u30eb\u30ad\u30e5\u30ea\u30a23 -\u8d64\u304d\u904b\u547d\u306e\u6226\u4e59\u5973- , lit . Valkyria of the Battlefield 3 -The Valkyrie of the Crimson Fate ) , illustrated by Mizuki Tsuge and eventually released in a single volume by Kadokawa Shoten in 2012 . \n",
|
| 92 |
+
" N @-@ 88 was unofficially designated around 1937 , connecting from N @-@ 29 , to N @-@ 86 and N @-@ 19 in Bridgeport . The route remained relatively the same as the state highway system was officially designated . Before 1955 , Nebraska did not have an adequate legal instrument to define the state highway system . By 1960 , N @-@ 19 was renumbered to US 385 , and US 26 was rerouted north near Bridgeport . The old alignment became part of N @-@ 92 . Two years later , N @-@ 29 was renumbered to N @-@ 71 . Between 1981 @-@ 82 , a road appeared on the official state map , extending from WYO 151 to N @-@ 71 . That road became part of N @-@ 88 by 1986 . No significant changes have been made since . \n",
|
| 93 |
+
" Assists ; Pts \n",
|
| 94 |
+
" Lively Numbers ; Macmillan , 1957 \n",
|
| 95 |
+
" St. John the Baptist ; central banner panel , Abesford church , 1943 \n",
|
| 96 |
+
" Lettering , sword , and shield ; mount for a list of men and woman serving in the Forces , St. Andrews , Croydon , 1943 \n",
|
| 97 |
+
" The tower had two floors and a mezzanine . On the first floor was the famous Sundial Room or Meridian Room , which was initially an open loggia . Pope Urban VIII had it enclosed and it was subsequently decorated with long sequences of frescoes painted between 1580 and 1582 by Simon Lagi and the two Flemish artists Paul and Matthijs Bril . Today the tower has paintings by Cristoforo Roncalli and Matteino da Siena . \n",
|
| 98 |
+
" Cicely Mary Barker ( 28 June 1895 \u2013 16 February 1973 ) was an English illustrator best known for a series of fantasy illustrations depicting fairies and flowers . Barker 's art education began in girlhood with correspondence courses and instruction at the Croydon School of Art . Her earliest professional work included greeting cards and juvenile magazine illustrations , and her first book , Flower Fairies of the Spring , was published in 1923 . Similar books were published in the following decades . \n",
|
| 99 |
+
" = = Description = = \n",
|
| 100 |
+
" Lt. Col. Dunnington continued to build up his works at Little Rock until November 1862 , when Captain Sanford C. Faulkner ( composer of The Arkansas Traveler ) was placed in charge of the Arsenal . Dunnington presumably returned to his naval duties and the Ponchartrain . \n",
|
| 101 |
+
" When Spring Came In at the Window ; Blackie , 1942 \n",
|
| 102 |
+
" Flower Fairies of the Wayside ; Blackie , 1948 \n",
|
| 103 |
+
" Losses ; OT \n",
|
| 104 |
+
" = = = Early life = = = \n",
|
| 105 |
+
" = = Biography = = \n",
|
| 106 |
+
" = = Legacy = = \n",
|
| 107 |
+
" Out of Great Tribulation ; memorial chapel , Norbury Medthodist church , 1948 \n",
|
| 108 |
+
" = Position ; GP = \n",
|
| 109 |
+
" Passed over in the Entry Draft , Cullen was finally selected by the Buffalo Sabres in the 1986 NHL Supplemental Draft . When the Sabres failed to offer him a contract , Cullen signed with the Flint Spirits of the International Hockey League ( IHL ) for the 1987 \u2013 88 season . He led the league with 157 points , scoring 48 goals , and won the James Gatschene Memorial Trophy as league most valuable player while sharing the Gary F. Longman Memorial Trophy with Ed Belfour as rookie of the year . Cullen 's outstanding season in Flint caught the attention of the Sabres and the Pittsburgh Penguins . He signed a contract with the Penguins for the league minimum , passing up a superior contract offer from Buffalo as he remained upset at how they released him the year before . \n",
|
| 110 |
+
" At the same time , his mother Loretta was diagnosed with skin cancer . Following her death early in his freshman year , Cullen contemplated returning to his Ontario home , but was convinced by his father to continue with both school and hockey . He used the game to cope with the loss and dedicated every game he played to his mother 's memory . Cullen felt that the inspiration he drew from his mother 's battle allowed him to become a better player . \n",
|
| 111 |
+
" The game 's battle system , the BliTZ system , is carried over directly from Valkyira Chronicles . During missions , players select each unit using a top @-@ down perspective of the battlefield map : once a character is selected , the player moves the character around the battlefield in third @-@ person . A character can only act once per @-@ turn , but characters can be granted multiple turns at the expense of other characters ' turns . Each character has a field and distance of movement limited by their Action Gauge . Up to nine characters can be assigned to a single mission . During gameplay , characters will call out if something happens to them , such as their health points ( HP ) getting low or being knocked out by enemy attacks . Each character has specific \" Potentials \" , skills unique to each character . They are divided into \" Personal Potential \" , which are innate skills that remain unaltered unless otherwise dictated by the story and can either help or impede a character , and \" Battle Potentials \" , which are grown throughout the game and always grant boons to a character . To learn Battle Potentials , each character has a unique \" Masters Table \" , a grid @-@ based skill table that can be used to acquire and link different skills . Characters also have Special Abilities that grant them temporary boosts on the battlefield : Kurt can activate \" Direct Command \" and move around the battlefield without depleting his Action Point gauge , the character Reila can shift into her \" Valkyria Form \" and become invincible , while Imca can target multiple enemy units with her heavy weapon . \n",
|
| 112 |
+
" = = The team = = \n",
|
| 113 |
+
" The military force at Little Rock under Dunnington 's command consisted of four officers : himself , Major John B. Lockman , Captain C.C. Green , and 2nd Lt. W.W. Murphy . In addition to these , he had 20 enlisted men and a civilian force composed of a foreman , 2 clerks , 3 gunsmiths for repairing small arms , a laboratorian , 26 laborers in the ammunition laboratory , and a carpenter for making packing boxes . \n",
|
| 114 |
+
" = = Works = = \n",
|
| 115 |
+
" Repaired and painted : \n",
|
| 116 |
+
" 752 packages of ordnance and ordnance stores received and mostly issued to troops in service . \n",
|
| 117 |
+
" Unlike its two predecessors , Valkyria Chronicles III was not released in the west . According to Sega , this was due to poor sales of Valkyria Chronicles II and the general unpopularity of the PSP in the west . An unofficial fan translation patch began development in February 2012 : players with a copy of Valkyria Chronicles III could download and apply the patch , which translated the game 's text into English . Compatible with the Extra Edition , the patch was released in January 2014 . \n",
|
| 118 |
+
" Shakespeare 's Girl Characters ; C. W. Faulkner , 1920 \n",
|
| 119 |
+
" The music was composed by Hitoshi Sakimoto , who had also worked on the previous Valkyria Chronicles games . When he originally heard about the project , he thought it would be a light tone similar to other Valkyria Chronicles games , but found the themes much darker than expected . An early theme he designed around his original vision of the project was rejected . He redid the main theme about seven times through the music production due to this need to reassess the game . The main theme was initially recorded using orchestra , then Sakimoto removed elements such as the guitar and bass , then adjusted the theme using a synthesizer before redoing segments such as the guitar piece on their own before incorporating them into the theme . The rejected main theme was used as a hopeful tune that played during the game 's ending . The battle themes were designed around the concept of a \" modern battle \" divorced from a fantasy scenario by using modern musical instruments , constructed to create a sense of atonality . While Sakimoto was most used to working with synthesized music , he felt that he needed to incorporate live instruments such as orchestra and guitar . The guitar was played by Mitsuhiro Ohta , who also arranged several of the later tracks . The game 's opening theme song , \" If You Wish for ... \" ( \u3082\u3057\u3082\u541b\u304c\u9858\u3046\u306e\u306a\u3089 , Moshimo Kimi ga Negauno Nara ) , was sung by Japanese singer May 'n . Its theme was the reason soldiers fought , in particular their wish to protect what was precious to them rather than a sense of responsibility or duty . Its lyrics were written by Seiko Fujibayashi , who had worked on May 'n on previous singles . \n",
|
| 120 |
+
" The item was intended simply as a piece of news , but telegraph lines quickly spread the news throughout the state , fueling procession sentiment . The rumor was interpreted by some Arkansans as a call from the governor to assemble to help expel the federal troops from the arsenal . By February 5 , six militia units , consisting of 1 @,@ 000 men , with a guarantee that the numbers could be increased to 5 @,@ 000 if the situations deemed it necessary , had assembled in Little Rock . Governor Rector vehemently denied ordering the troops to assemble or giving any order at all in connection with the troops . Faced with the fact that the military had assembled believing they were following his orders and the consensus of the citizens of Little Rock against any armed conflict between the civilian army and federal troops , Governor Rector was forced to take control of the situation . On February 6 , he sent a formal demand for surrender of the arsenal to Captain Totten , \n",
|
| 121 |
+
" Fernandez shares a close bond with her family , and admits to missing being around them . She says : \" I miss them so much everyday . You don 't realise when you live away from home how difficult life can be [ ... ] At the same time , staying away from them has taught me to be more responsible . It has taught me so many things about myself , about priorities and time management . \" In March 2012 , Fernandez turned vegetarian for a 40 @-@ day period to observe Lent , a period from Ash Wednesday to Holy Saturday . \n",
|
| 122 |
+
" The team was involved in a controversial loss to the Los Angeles Kings , when the Staples Center clock appeared to freeze at 1 @.@ 8 seconds allowing the Kings time to score the tying goal , before winning in overtime . During the season Columbus managed only two winning streaks of three or more games . One of which came towards the end of the year helping the Blue Jackets finish with 65 points , the third worst point total in franchise history . \n",
|
| 123 |
+
" = Saves ; Sv % = \n",
|
| 124 |
+
" = = History = = \n",
|
| 125 |
+
" Famitsu enjoyed the story , and were particularly pleased with the improvements to gameplay . Japanese gaming site Game Watch Impress , despite negatively noting its pacing and elements recycled from previous games , was generally positive about its story and characters , and found its gameplay entertaining despite off @-@ putting difficulty spikes . 4Gamer.net writer Naohiko Misuosame , in a \" Play Test \" article based on the game 's PSN demo , felt that Valkyria Chronicles III provided a \" profound feeling of closure \" for the Valkyria Chronicles series . He praised its gameplay despite annoying limitations to aspects such as special abilities , and positively noted its shift in story to a tone similar to the first game . \n",
|
| 126 |
+
" Three days later , on March 28 , goaltender Steve Mason was injured in the morning skate when a shot from Colton Gillies hit him in the mask . With Sanford again injured , York made an emergency start . Playing against the Detroit Red Wings , York made 29 saves , including 17 in the third period , helping Columbus to a 4 \u2013 2 victory and giving York his first career NHL win . York remained the starter and led the Blue Jackets to a second three @-@ game winning streak . In his fourth start , Columbus was shutout by the Coyotes despite a franchise @-@ record 54 shots on goal , losing 2 \u2013 0 . The 54 saves by Phoenix goaltender Mike Smith set an NHL record for a regulation shutout . Mason returned to the starter 's role for the final two games , winning both . The two victories gave Columbus 65 points for the year , their third @-@ lowest total in franchise history . \n",
|
| 127 |
+
" The first stage of building of the tower , as recorded by Leo XIII in his motu proprio Ut mysticam of 1891 , is credited to Pope Gregory XIII , Pope from 1572 to 1585 . The directive was to build a tower at a suitable location in the Vatican and equip it with the \" greatest and best instruments of the time \" . The design was effected after a series of meetings of the experts who had been appointed to reform the Julian calendar , in use since 45 BC , to verify their proposed reforms . Fr . Christoph Clavius , a Jesuit mathematician from the Roman College , was the expert on the committee who suggested the new system for the observations . The 73 metres ( 240 ft ) tower was then built above the museum and library , flanked by the Belvedere and della Pigna courtyards . The instrumentation for the observation of the sun rays falling over it consisted of a meridian line designed by Ignazio Danti of Perugia . It was in the form of a circular marble plate in the centre , embellished with scientific designs . The tower still remains today , but has undergone improvements over the centuries . \n",
|
| 128 |
+
" In 1924 , the family moved into a four @-@ level , semi @-@ detached Victorian house at 23 The Waldrons . Barker had a studio built in the garden and her sister conducted a kindergarten in a room at the back of the house . The family lived frugally and attended both St. Edmund 's and St. Andrew 's in Croydon \u2013 \" low \" churches for the less privileged . Barker sometimes incorporated portraits of her fellow parishioners in her religious works . She was described by Canon Ingram Hill as \" one of the pillars \" of St. Andrew 's . \n",
|
| 129 |
+
" The Penguins ' needs led them to complete a blockbuster trade on March 1 , 1991 . Cullen was sent to the Hartford Whalers , along with Zarley Zalapski and Jeff Parker in exchange for Hartford 's all @-@ time leading scorer , Ron Francis , along with Ulf Samuelsson and Grant Jennings . The Penguins almost turned down the deal as they were concerned about giving up Cullen 's playmaking and leadership abilities , while his former teammates credited Cullen as being the primary reason they were in a playoff position at the time the trade happened . After the Penguins won their first Stanley Cup that season , Phil Bourque later said it \" broke his heart \" that Cullen was not able to share in that championship . \n",
|
| 130 |
+
" It met with positive sales in Japan , and was praised by both Japanese and western critics . After release , it received downloadable content , along with an expanded edition in November of that year . It was also adapted into manga and an original video animation series . Due to low sales of Valkyria Chronicles II , Valkyria Chronicles III was not localized , but a fan translation compatible with the game 's expanded edition was released in 2014 . Media.Vision would return to the franchise with the development of Valkyria : Azure Revolution for the PlayStation 4 . \n",
|
| 131 |
+
" = = = Christian @-@ themed works = = = \n",
|
| 132 |
+
" = Goals ; A = \n",
|
| 133 |
+
" Autumn Songs with Music ; Blackie , 1927 \n",
|
| 134 |
+
" Banner design ; St. Mary 's , Sanderstead , 1923 \n",
|
| 135 |
+
" Save Percentage ; SO = Shutouts \n",
|
| 136 |
+
" = = = Goaltenders = = = \n",
|
| 137 |
+
" The \" Summaries of Work done at Little Rock Arsenal , C.S.A. \" continue at about the same pace and scale from August 1862 until August 1863 . Appended to the \" Summary \" for August , 1863 is the ominous notation , \" During the last week in the month , nearly all stores at the Arsenal have been packed and sent to Arkadelphia , in obedience to orders from Chief of Ordnance , District of Arkansas . \" This then marks the beginning of the evacuation of ordnance activities from Little Rock , with the city being surrendered to the advancing Federal troops of Frederick Steele 's Arkansas Expedition on September 11 , 1863 . \n",
|
| 138 |
+
" = Tower Building of the Little Rock Arsenal = \n",
|
| 139 |
+
" This ammunition , and that which I brought with me , was rapidly prepared for use at the Laboratory established at the Little Rock Arsenal for that purpose . As illustrating as the pitiful scarcity of material in the country , the fact may be stated that it was found necessary to use public documents of the State Library for cartridge paper . Gunsmiths were employed or conscripted , tools purchased or impressed , and the repair of the damaged guns I brought with me and about an equal number found at Little Rock commenced at once . But , after inspecting the work and observing the spirit of the men I decided that a garrison 500 strong could hold out against Fitch and that I would lead the remainder - about 1500 - to Gen 'l Rust as soon as shotguns and rifles could be obtained from Little Rock instead of pikes and lances , with which most of them were armed . Two days elapsed before the change could be effected . \" \n",
|
| 140 |
+
" The building and the surrounding park were used for many public purposes throughout the early 20th century . The Tower Building served as headquarters for the United Confederate Veterans Reunion , May 15 \u2013 18 , 1911 . Over 106 @,@ 000 Civil War veterans , the largest popular gathering in the history of the city up to that time , attended and were housed in the building or camped in the park , which had also become a popular camping area . Later the building served as an armory for the Arkansas National Guard . In 1912 , the second floor of the Tower Building became Little Rock 's first public library . In 1917 , Little Rock built a fire station in the park , that building is now gone . A band shell named for H. H. Foster also was built in the park during this time , but also no longer exists . In 1936 , Works Progress Administration built the Museum of Fine Arts , now called the Arkansas Arts Center , just south of the Tower Building . \n",
|
| 141 |
+
" Following the break , the Blue Jackets were on the road playing the Los Angeles Kings , and with the score tied late in the game , Kings ' defenseman Drew Doughty scored with just 0 @.@ 4 seconds remaining to win the game . Upon review of the goal it , was determined that the clock at Staples Center froze at 1 @.@ 8 seconds for over a full second , which would have resulted in time expiring prior to the goal being scored . Kings ' General Manager Dean Lombardi stated that the clock was correct and no extra time had been added due to the way the clock self @-@ corrects at various times . Howson stated on the team 's blog that \" It is an amazing coincidence that with the Kings on a power play at Staples Center and with a mad scramble around our net in the dying seconds of the third period of a 2 \u2013 2 hockey game that the clock stopped for at least one full second , \" adding that , \" Either there was a deliberate stopping of the clock or the clock malfunctioned . \" NHL Senior Vice President of Hockey Operations Colin Campbell stated that the Blue Jackets were wronged , but that the outcome of the game could not be changed , and that the delay was not noticed by the off @-@ ice officials or the situation room in Toronto . To determine the true cause of the clock pause , the NHL launched an investigation , talking with the clock 's manufacturer and interviewing Staples Center staff . \n",
|
| 142 |
+
" On 31 July 1862 , under the command of Lieutenant Charles H. McBlair , Atlanta conducted her sea trials down the Savannah River toward Fort Pulaski . The ship proved to be difficult to steer , and the additional weight of her armor and guns significantly reduced her speed and increased her draft . This latter was a real problem in the shallow waters near Savannah . She also leaked significantly , and her design virtually eliminated air circulation . One report said that \" it was almost intolerable on board the Atlanta , there being no method of ventilation , and the heat was intense . \" Scales commented in his diary , \" What a comfortless , infernal and God @-@ forsaken ship ! ! \" \n",
|
| 143 |
+
" = = Human interactions = = \n",
|
| 144 |
+
" In 2014 , Fernandez appeared in Sajid Nadiadwala 's directorial debut \u2014 the action film Kick , a remake of a 2009 Telugu film of same name . She starred opposite Salman Khan , playing Shaina , a psychiatry student . She retained her real voice for the first time in Kick . While Sneha May Francis commented that she is : \" incredibly dazzling , and moves like a magic \" , Raja Sen of Rediff.com was more critical of her dialogue delivery , calling it \" unfortunate . \" The film received mixed reviews from critics , but with worldwide revenue of over \u20b9 3 @.@ 75 billion ( US $ 56 million ) , it became the fourth highest @-@ grossing Bollywood film . The film established Fernandez as one of the most popular Bollywood actresses . \n",
|
| 145 |
+
" A Flower Fairies Treasury ; Frederick Warne , 1997 \n",
|
| 146 |
+
" M1855 .58 cal rifle @-@ muskets 900 \n",
|
| 147 |
+
" The armor of the casemate was angled at 30 \u00b0 from the horizontal and made from two layers of railroad rails , rolled into plates 2 inches ( 51 mm ) thick and 7 inches ( 180 mm ) wide . The outer layer ran vertically and the inner layer horizontally . Her armor was backed by 3 inches ( 76 mm ) of oak , vertically oriented , and two layers of 7 @.@ 5 inches ( 191 mm ) of pine , alternating in direction . The bottom of the casemate was some 20 inches ( 508 mm ) from the waterline and its top was 8 feet 6 inches ( 2 @.@ 59 m ) above the waterline . The pyramidal pilothouse was armored in the same way and had room for two men . The upper portion of Atlanta 's hull received two inches of armor . \n",
|
| 148 |
+
" Baptismal rolls ; St. Andrews , Croydon , 1948 , 1962 \n",
|
| 149 |
+
" The family of four was moderately well off , and belonged to the lower end of the upper middle class . A nanny , a governess , and a cook to prepare special meals for Barker were hired . She spent much time in bed at home amusing herself with painting books and a nursery library that included the works of Kate Greenaway and Randolph Caldecott \u2013 two artists who exerted strong influences on her later art . \n",
|
| 150 |
+
" \u2020 Denotes player spent time with another team before joining Blue Jackets . Stats reflect time with the Blue Jackets only . \u2021 Traded mid @-@ season \n",
|
| 151 |
+
" = = = Depictions of children = = = \n",
|
| 152 |
+
" Total 10 @,@ 247 \n",
|
| 153 |
+
" = = Taxonomy and phylogeny = = \n",
|
| 154 |
+
" Barker took correspondence courses in art , probably until about 1919 . In 1908 at 13 years , she entered an evening class at the Croydon School of Art , and attended the school into the 1940s . In time , she received a teaching position . \n"
|
| 155 |
+
],
|
| 156 |
+
"desc_act": false,
|
| 157 |
+
"exllama_config": {
|
| 158 |
+
"version": 1
|
| 159 |
+
},
|
| 160 |
+
"group_size": 128,
|
| 161 |
+
"max_input_length": null,
|
| 162 |
+
"model_seqlen": null,
|
| 163 |
+
"module_name_preceding_first_block": null,
|
| 164 |
+
"modules_in_block_to_quantize": null,
|
| 165 |
+
"pad_token_id": null,
|
| 166 |
+
"quant_method": "gptq",
|
| 167 |
+
"sym": true,
|
| 168 |
+
"tokenizer": null,
|
| 169 |
+
"true_sequential": true,
|
| 170 |
+
"use_cuda_fp16": false,
|
| 171 |
+
"use_exllama": true
|
| 172 |
+
},
|
| 173 |
+
"rms_norm_eps": 1e-05,
|
| 174 |
+
"rope_theta": 1000000.0,
|
| 175 |
+
"sliding_window": null,
|
| 176 |
+
"tie_word_embeddings": false,
|
| 177 |
+
"torch_dtype": "float16",
|
| 178 |
+
"transformers_version": "4.47.1",
|
| 179 |
+
"unsloth_version": "2024.8",
|
| 180 |
+
"use_cache": true,
|
| 181 |
+
"vocab_size": 131072
|
| 182 |
+
}
|
generation_config.json
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_from_model_config": true,
|
| 3 |
+
"bos_token_id": 1,
|
| 4 |
+
"eos_token_id": 2,
|
| 5 |
+
"pad_token_id": 10,
|
| 6 |
+
"transformers_version": "4.47.1"
|
| 7 |
+
}
|
model-00001-of-00003.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:db57608fc8fdd14a243ce7523e1f56a49047e5dca4a8fed365aa7aac984a23c4
|
| 3 |
+
size 4998897472
|
model-00002-of-00003.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:659e3347285213a0a1d58fac4a92ad7a737b3a499ab42321e5ad320615259601
|
| 3 |
+
size 4999084904
|
model-00003-of-00003.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f8f698505d6a39800db7600d632df348d7ae4543595d2461e6da35b89934b730
|
| 3 |
+
size 3855163384
|
model.safetensors.index.json
ADDED
|
@@ -0,0 +1,1210 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"metadata": {
|
| 3 |
+
"total_size": 13853009920
|
| 4 |
+
},
|
| 5 |
+
"weight_map": {
|
| 6 |
+
"lm_head.weight": "model-00003-of-00003.safetensors",
|
| 7 |
+
"model.embed_tokens.weight": "model-00001-of-00003.safetensors",
|
| 8 |
+
"model.layers.0.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 9 |
+
"model.layers.0.mlp.down_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 10 |
+
"model.layers.0.mlp.down_proj.qweight": "model-00001-of-00003.safetensors",
|
| 11 |
+
"model.layers.0.mlp.down_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 12 |
+
"model.layers.0.mlp.down_proj.scales": "model-00001-of-00003.safetensors",
|
| 13 |
+
"model.layers.0.mlp.gate_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 14 |
+
"model.layers.0.mlp.gate_proj.qweight": "model-00001-of-00003.safetensors",
|
| 15 |
+
"model.layers.0.mlp.gate_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 16 |
+
"model.layers.0.mlp.gate_proj.scales": "model-00001-of-00003.safetensors",
|
| 17 |
+
"model.layers.0.mlp.up_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 18 |
+
"model.layers.0.mlp.up_proj.qweight": "model-00001-of-00003.safetensors",
|
| 19 |
+
"model.layers.0.mlp.up_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 20 |
+
"model.layers.0.mlp.up_proj.scales": "model-00001-of-00003.safetensors",
|
| 21 |
+
"model.layers.0.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 22 |
+
"model.layers.0.self_attn.k_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 23 |
+
"model.layers.0.self_attn.k_proj.qweight": "model-00001-of-00003.safetensors",
|
| 24 |
+
"model.layers.0.self_attn.k_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 25 |
+
"model.layers.0.self_attn.k_proj.scales": "model-00001-of-00003.safetensors",
|
| 26 |
+
"model.layers.0.self_attn.o_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 27 |
+
"model.layers.0.self_attn.o_proj.qweight": "model-00001-of-00003.safetensors",
|
| 28 |
+
"model.layers.0.self_attn.o_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 29 |
+
"model.layers.0.self_attn.o_proj.scales": "model-00001-of-00003.safetensors",
|
| 30 |
+
"model.layers.0.self_attn.q_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 31 |
+
"model.layers.0.self_attn.q_proj.qweight": "model-00001-of-00003.safetensors",
|
| 32 |
+
"model.layers.0.self_attn.q_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 33 |
+
"model.layers.0.self_attn.q_proj.scales": "model-00001-of-00003.safetensors",
|
| 34 |
+
"model.layers.0.self_attn.v_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 35 |
+
"model.layers.0.self_attn.v_proj.qweight": "model-00001-of-00003.safetensors",
|
| 36 |
+
"model.layers.0.self_attn.v_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 37 |
+
"model.layers.0.self_attn.v_proj.scales": "model-00001-of-00003.safetensors",
|
| 38 |
+
"model.layers.1.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 39 |
+
"model.layers.1.mlp.down_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 40 |
+
"model.layers.1.mlp.down_proj.qweight": "model-00001-of-00003.safetensors",
|
| 41 |
+
"model.layers.1.mlp.down_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 42 |
+
"model.layers.1.mlp.down_proj.scales": "model-00001-of-00003.safetensors",
|
| 43 |
+
"model.layers.1.mlp.gate_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 44 |
+
"model.layers.1.mlp.gate_proj.qweight": "model-00001-of-00003.safetensors",
|
| 45 |
+
"model.layers.1.mlp.gate_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 46 |
+
"model.layers.1.mlp.gate_proj.scales": "model-00001-of-00003.safetensors",
|
| 47 |
+
"model.layers.1.mlp.up_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 48 |
+
"model.layers.1.mlp.up_proj.qweight": "model-00001-of-00003.safetensors",
|
| 49 |
+
"model.layers.1.mlp.up_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 50 |
+
"model.layers.1.mlp.up_proj.scales": "model-00001-of-00003.safetensors",
|
| 51 |
+
"model.layers.1.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 52 |
+
"model.layers.1.self_attn.k_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 53 |
+
"model.layers.1.self_attn.k_proj.qweight": "model-00001-of-00003.safetensors",
|
| 54 |
+
"model.layers.1.self_attn.k_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 55 |
+
"model.layers.1.self_attn.k_proj.scales": "model-00001-of-00003.safetensors",
|
| 56 |
+
"model.layers.1.self_attn.o_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 57 |
+
"model.layers.1.self_attn.o_proj.qweight": "model-00001-of-00003.safetensors",
|
| 58 |
+
"model.layers.1.self_attn.o_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 59 |
+
"model.layers.1.self_attn.o_proj.scales": "model-00001-of-00003.safetensors",
|
| 60 |
+
"model.layers.1.self_attn.q_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 61 |
+
"model.layers.1.self_attn.q_proj.qweight": "model-00001-of-00003.safetensors",
|
| 62 |
+
"model.layers.1.self_attn.q_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 63 |
+
"model.layers.1.self_attn.q_proj.scales": "model-00001-of-00003.safetensors",
|
| 64 |
+
"model.layers.1.self_attn.v_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 65 |
+
"model.layers.1.self_attn.v_proj.qweight": "model-00001-of-00003.safetensors",
|
| 66 |
+
"model.layers.1.self_attn.v_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 67 |
+
"model.layers.1.self_attn.v_proj.scales": "model-00001-of-00003.safetensors",
|
| 68 |
+
"model.layers.10.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 69 |
+
"model.layers.10.mlp.down_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 70 |
+
"model.layers.10.mlp.down_proj.qweight": "model-00001-of-00003.safetensors",
|
| 71 |
+
"model.layers.10.mlp.down_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 72 |
+
"model.layers.10.mlp.down_proj.scales": "model-00001-of-00003.safetensors",
|
| 73 |
+
"model.layers.10.mlp.gate_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 74 |
+
"model.layers.10.mlp.gate_proj.qweight": "model-00001-of-00003.safetensors",
|
| 75 |
+
"model.layers.10.mlp.gate_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 76 |
+
"model.layers.10.mlp.gate_proj.scales": "model-00001-of-00003.safetensors",
|
| 77 |
+
"model.layers.10.mlp.up_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 78 |
+
"model.layers.10.mlp.up_proj.qweight": "model-00001-of-00003.safetensors",
|
| 79 |
+
"model.layers.10.mlp.up_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 80 |
+
"model.layers.10.mlp.up_proj.scales": "model-00001-of-00003.safetensors",
|
| 81 |
+
"model.layers.10.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 82 |
+
"model.layers.10.self_attn.k_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 83 |
+
"model.layers.10.self_attn.k_proj.qweight": "model-00001-of-00003.safetensors",
|
| 84 |
+
"model.layers.10.self_attn.k_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 85 |
+
"model.layers.10.self_attn.k_proj.scales": "model-00001-of-00003.safetensors",
|
| 86 |
+
"model.layers.10.self_attn.o_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 87 |
+
"model.layers.10.self_attn.o_proj.qweight": "model-00001-of-00003.safetensors",
|
| 88 |
+
"model.layers.10.self_attn.o_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 89 |
+
"model.layers.10.self_attn.o_proj.scales": "model-00001-of-00003.safetensors",
|
| 90 |
+
"model.layers.10.self_attn.q_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 91 |
+
"model.layers.10.self_attn.q_proj.qweight": "model-00001-of-00003.safetensors",
|
| 92 |
+
"model.layers.10.self_attn.q_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 93 |
+
"model.layers.10.self_attn.q_proj.scales": "model-00001-of-00003.safetensors",
|
| 94 |
+
"model.layers.10.self_attn.v_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 95 |
+
"model.layers.10.self_attn.v_proj.qweight": "model-00001-of-00003.safetensors",
|
| 96 |
+
"model.layers.10.self_attn.v_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 97 |
+
"model.layers.10.self_attn.v_proj.scales": "model-00001-of-00003.safetensors",
|
| 98 |
+
"model.layers.11.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 99 |
+
"model.layers.11.mlp.down_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 100 |
+
"model.layers.11.mlp.down_proj.qweight": "model-00001-of-00003.safetensors",
|
| 101 |
+
"model.layers.11.mlp.down_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 102 |
+
"model.layers.11.mlp.down_proj.scales": "model-00001-of-00003.safetensors",
|
| 103 |
+
"model.layers.11.mlp.gate_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 104 |
+
"model.layers.11.mlp.gate_proj.qweight": "model-00001-of-00003.safetensors",
|
| 105 |
+
"model.layers.11.mlp.gate_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 106 |
+
"model.layers.11.mlp.gate_proj.scales": "model-00001-of-00003.safetensors",
|
| 107 |
+
"model.layers.11.mlp.up_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 108 |
+
"model.layers.11.mlp.up_proj.qweight": "model-00001-of-00003.safetensors",
|
| 109 |
+
"model.layers.11.mlp.up_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 110 |
+
"model.layers.11.mlp.up_proj.scales": "model-00001-of-00003.safetensors",
|
| 111 |
+
"model.layers.11.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 112 |
+
"model.layers.11.self_attn.k_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 113 |
+
"model.layers.11.self_attn.k_proj.qweight": "model-00001-of-00003.safetensors",
|
| 114 |
+
"model.layers.11.self_attn.k_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 115 |
+
"model.layers.11.self_attn.k_proj.scales": "model-00001-of-00003.safetensors",
|
| 116 |
+
"model.layers.11.self_attn.o_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 117 |
+
"model.layers.11.self_attn.o_proj.qweight": "model-00001-of-00003.safetensors",
|
| 118 |
+
"model.layers.11.self_attn.o_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 119 |
+
"model.layers.11.self_attn.o_proj.scales": "model-00001-of-00003.safetensors",
|
| 120 |
+
"model.layers.11.self_attn.q_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 121 |
+
"model.layers.11.self_attn.q_proj.qweight": "model-00001-of-00003.safetensors",
|
| 122 |
+
"model.layers.11.self_attn.q_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 123 |
+
"model.layers.11.self_attn.q_proj.scales": "model-00001-of-00003.safetensors",
|
| 124 |
+
"model.layers.11.self_attn.v_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 125 |
+
"model.layers.11.self_attn.v_proj.qweight": "model-00001-of-00003.safetensors",
|
| 126 |
+
"model.layers.11.self_attn.v_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 127 |
+
"model.layers.11.self_attn.v_proj.scales": "model-00001-of-00003.safetensors",
|
| 128 |
+
"model.layers.12.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 129 |
+
"model.layers.12.mlp.down_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 130 |
+
"model.layers.12.mlp.down_proj.qweight": "model-00001-of-00003.safetensors",
|
| 131 |
+
"model.layers.12.mlp.down_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 132 |
+
"model.layers.12.mlp.down_proj.scales": "model-00001-of-00003.safetensors",
|
| 133 |
+
"model.layers.12.mlp.gate_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 134 |
+
"model.layers.12.mlp.gate_proj.qweight": "model-00001-of-00003.safetensors",
|
| 135 |
+
"model.layers.12.mlp.gate_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 136 |
+
"model.layers.12.mlp.gate_proj.scales": "model-00001-of-00003.safetensors",
|
| 137 |
+
"model.layers.12.mlp.up_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 138 |
+
"model.layers.12.mlp.up_proj.qweight": "model-00001-of-00003.safetensors",
|
| 139 |
+
"model.layers.12.mlp.up_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 140 |
+
"model.layers.12.mlp.up_proj.scales": "model-00001-of-00003.safetensors",
|
| 141 |
+
"model.layers.12.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 142 |
+
"model.layers.12.self_attn.k_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 143 |
+
"model.layers.12.self_attn.k_proj.qweight": "model-00001-of-00003.safetensors",
|
| 144 |
+
"model.layers.12.self_attn.k_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 145 |
+
"model.layers.12.self_attn.k_proj.scales": "model-00001-of-00003.safetensors",
|
| 146 |
+
"model.layers.12.self_attn.o_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 147 |
+
"model.layers.12.self_attn.o_proj.qweight": "model-00001-of-00003.safetensors",
|
| 148 |
+
"model.layers.12.self_attn.o_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 149 |
+
"model.layers.12.self_attn.o_proj.scales": "model-00001-of-00003.safetensors",
|
| 150 |
+
"model.layers.12.self_attn.q_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 151 |
+
"model.layers.12.self_attn.q_proj.qweight": "model-00001-of-00003.safetensors",
|
| 152 |
+
"model.layers.12.self_attn.q_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 153 |
+
"model.layers.12.self_attn.q_proj.scales": "model-00001-of-00003.safetensors",
|
| 154 |
+
"model.layers.12.self_attn.v_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 155 |
+
"model.layers.12.self_attn.v_proj.qweight": "model-00001-of-00003.safetensors",
|
| 156 |
+
"model.layers.12.self_attn.v_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 157 |
+
"model.layers.12.self_attn.v_proj.scales": "model-00001-of-00003.safetensors",
|
| 158 |
+
"model.layers.13.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 159 |
+
"model.layers.13.mlp.down_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 160 |
+
"model.layers.13.mlp.down_proj.qweight": "model-00002-of-00003.safetensors",
|
| 161 |
+
"model.layers.13.mlp.down_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 162 |
+
"model.layers.13.mlp.down_proj.scales": "model-00002-of-00003.safetensors",
|
| 163 |
+
"model.layers.13.mlp.gate_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 164 |
+
"model.layers.13.mlp.gate_proj.qweight": "model-00002-of-00003.safetensors",
|
| 165 |
+
"model.layers.13.mlp.gate_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 166 |
+
"model.layers.13.mlp.gate_proj.scales": "model-00002-of-00003.safetensors",
|
| 167 |
+
"model.layers.13.mlp.up_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 168 |
+
"model.layers.13.mlp.up_proj.qweight": "model-00002-of-00003.safetensors",
|
| 169 |
+
"model.layers.13.mlp.up_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 170 |
+
"model.layers.13.mlp.up_proj.scales": "model-00002-of-00003.safetensors",
|
| 171 |
+
"model.layers.13.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 172 |
+
"model.layers.13.self_attn.k_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 173 |
+
"model.layers.13.self_attn.k_proj.qweight": "model-00001-of-00003.safetensors",
|
| 174 |
+
"model.layers.13.self_attn.k_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 175 |
+
"model.layers.13.self_attn.k_proj.scales": "model-00001-of-00003.safetensors",
|
| 176 |
+
"model.layers.13.self_attn.o_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 177 |
+
"model.layers.13.self_attn.o_proj.qweight": "model-00001-of-00003.safetensors",
|
| 178 |
+
"model.layers.13.self_attn.o_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 179 |
+
"model.layers.13.self_attn.o_proj.scales": "model-00001-of-00003.safetensors",
|
| 180 |
+
"model.layers.13.self_attn.q_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 181 |
+
"model.layers.13.self_attn.q_proj.qweight": "model-00002-of-00003.safetensors",
|
| 182 |
+
"model.layers.13.self_attn.q_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 183 |
+
"model.layers.13.self_attn.q_proj.scales": "model-00002-of-00003.safetensors",
|
| 184 |
+
"model.layers.13.self_attn.v_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 185 |
+
"model.layers.13.self_attn.v_proj.qweight": "model-00002-of-00003.safetensors",
|
| 186 |
+
"model.layers.13.self_attn.v_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 187 |
+
"model.layers.13.self_attn.v_proj.scales": "model-00002-of-00003.safetensors",
|
| 188 |
+
"model.layers.14.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 189 |
+
"model.layers.14.mlp.down_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 190 |
+
"model.layers.14.mlp.down_proj.qweight": "model-00002-of-00003.safetensors",
|
| 191 |
+
"model.layers.14.mlp.down_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 192 |
+
"model.layers.14.mlp.down_proj.scales": "model-00002-of-00003.safetensors",
|
| 193 |
+
"model.layers.14.mlp.gate_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 194 |
+
"model.layers.14.mlp.gate_proj.qweight": "model-00002-of-00003.safetensors",
|
| 195 |
+
"model.layers.14.mlp.gate_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 196 |
+
"model.layers.14.mlp.gate_proj.scales": "model-00002-of-00003.safetensors",
|
| 197 |
+
"model.layers.14.mlp.up_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 198 |
+
"model.layers.14.mlp.up_proj.qweight": "model-00002-of-00003.safetensors",
|
| 199 |
+
"model.layers.14.mlp.up_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 200 |
+
"model.layers.14.mlp.up_proj.scales": "model-00002-of-00003.safetensors",
|
| 201 |
+
"model.layers.14.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 202 |
+
"model.layers.14.self_attn.k_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 203 |
+
"model.layers.14.self_attn.k_proj.qweight": "model-00002-of-00003.safetensors",
|
| 204 |
+
"model.layers.14.self_attn.k_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 205 |
+
"model.layers.14.self_attn.k_proj.scales": "model-00002-of-00003.safetensors",
|
| 206 |
+
"model.layers.14.self_attn.o_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 207 |
+
"model.layers.14.self_attn.o_proj.qweight": "model-00002-of-00003.safetensors",
|
| 208 |
+
"model.layers.14.self_attn.o_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 209 |
+
"model.layers.14.self_attn.o_proj.scales": "model-00002-of-00003.safetensors",
|
| 210 |
+
"model.layers.14.self_attn.q_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 211 |
+
"model.layers.14.self_attn.q_proj.qweight": "model-00002-of-00003.safetensors",
|
| 212 |
+
"model.layers.14.self_attn.q_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 213 |
+
"model.layers.14.self_attn.q_proj.scales": "model-00002-of-00003.safetensors",
|
| 214 |
+
"model.layers.14.self_attn.v_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 215 |
+
"model.layers.14.self_attn.v_proj.qweight": "model-00002-of-00003.safetensors",
|
| 216 |
+
"model.layers.14.self_attn.v_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 217 |
+
"model.layers.14.self_attn.v_proj.scales": "model-00002-of-00003.safetensors",
|
| 218 |
+
"model.layers.15.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 219 |
+
"model.layers.15.mlp.down_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 220 |
+
"model.layers.15.mlp.down_proj.qweight": "model-00002-of-00003.safetensors",
|
| 221 |
+
"model.layers.15.mlp.down_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 222 |
+
"model.layers.15.mlp.down_proj.scales": "model-00002-of-00003.safetensors",
|
| 223 |
+
"model.layers.15.mlp.gate_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 224 |
+
"model.layers.15.mlp.gate_proj.qweight": "model-00002-of-00003.safetensors",
|
| 225 |
+
"model.layers.15.mlp.gate_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 226 |
+
"model.layers.15.mlp.gate_proj.scales": "model-00002-of-00003.safetensors",
|
| 227 |
+
"model.layers.15.mlp.up_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 228 |
+
"model.layers.15.mlp.up_proj.qweight": "model-00002-of-00003.safetensors",
|
| 229 |
+
"model.layers.15.mlp.up_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 230 |
+
"model.layers.15.mlp.up_proj.scales": "model-00002-of-00003.safetensors",
|
| 231 |
+
"model.layers.15.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 232 |
+
"model.layers.15.self_attn.k_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 233 |
+
"model.layers.15.self_attn.k_proj.qweight": "model-00002-of-00003.safetensors",
|
| 234 |
+
"model.layers.15.self_attn.k_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 235 |
+
"model.layers.15.self_attn.k_proj.scales": "model-00002-of-00003.safetensors",
|
| 236 |
+
"model.layers.15.self_attn.o_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 237 |
+
"model.layers.15.self_attn.o_proj.qweight": "model-00002-of-00003.safetensors",
|
| 238 |
+
"model.layers.15.self_attn.o_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 239 |
+
"model.layers.15.self_attn.o_proj.scales": "model-00002-of-00003.safetensors",
|
| 240 |
+
"model.layers.15.self_attn.q_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 241 |
+
"model.layers.15.self_attn.q_proj.qweight": "model-00002-of-00003.safetensors",
|
| 242 |
+
"model.layers.15.self_attn.q_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 243 |
+
"model.layers.15.self_attn.q_proj.scales": "model-00002-of-00003.safetensors",
|
| 244 |
+
"model.layers.15.self_attn.v_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 245 |
+
"model.layers.15.self_attn.v_proj.qweight": "model-00002-of-00003.safetensors",
|
| 246 |
+
"model.layers.15.self_attn.v_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 247 |
+
"model.layers.15.self_attn.v_proj.scales": "model-00002-of-00003.safetensors",
|
| 248 |
+
"model.layers.16.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 249 |
+
"model.layers.16.mlp.down_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 250 |
+
"model.layers.16.mlp.down_proj.qweight": "model-00002-of-00003.safetensors",
|
| 251 |
+
"model.layers.16.mlp.down_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 252 |
+
"model.layers.16.mlp.down_proj.scales": "model-00002-of-00003.safetensors",
|
| 253 |
+
"model.layers.16.mlp.gate_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 254 |
+
"model.layers.16.mlp.gate_proj.qweight": "model-00002-of-00003.safetensors",
|
| 255 |
+
"model.layers.16.mlp.gate_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 256 |
+
"model.layers.16.mlp.gate_proj.scales": "model-00002-of-00003.safetensors",
|
| 257 |
+
"model.layers.16.mlp.up_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 258 |
+
"model.layers.16.mlp.up_proj.qweight": "model-00002-of-00003.safetensors",
|
| 259 |
+
"model.layers.16.mlp.up_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 260 |
+
"model.layers.16.mlp.up_proj.scales": "model-00002-of-00003.safetensors",
|
| 261 |
+
"model.layers.16.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 262 |
+
"model.layers.16.self_attn.k_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 263 |
+
"model.layers.16.self_attn.k_proj.qweight": "model-00002-of-00003.safetensors",
|
| 264 |
+
"model.layers.16.self_attn.k_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 265 |
+
"model.layers.16.self_attn.k_proj.scales": "model-00002-of-00003.safetensors",
|
| 266 |
+
"model.layers.16.self_attn.o_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 267 |
+
"model.layers.16.self_attn.o_proj.qweight": "model-00002-of-00003.safetensors",
|
| 268 |
+
"model.layers.16.self_attn.o_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 269 |
+
"model.layers.16.self_attn.o_proj.scales": "model-00002-of-00003.safetensors",
|
| 270 |
+
"model.layers.16.self_attn.q_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 271 |
+
"model.layers.16.self_attn.q_proj.qweight": "model-00002-of-00003.safetensors",
|
| 272 |
+
"model.layers.16.self_attn.q_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 273 |
+
"model.layers.16.self_attn.q_proj.scales": "model-00002-of-00003.safetensors",
|
| 274 |
+
"model.layers.16.self_attn.v_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 275 |
+
"model.layers.16.self_attn.v_proj.qweight": "model-00002-of-00003.safetensors",
|
| 276 |
+
"model.layers.16.self_attn.v_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 277 |
+
"model.layers.16.self_attn.v_proj.scales": "model-00002-of-00003.safetensors",
|
| 278 |
+
"model.layers.17.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 279 |
+
"model.layers.17.mlp.down_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 280 |
+
"model.layers.17.mlp.down_proj.qweight": "model-00002-of-00003.safetensors",
|
| 281 |
+
"model.layers.17.mlp.down_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 282 |
+
"model.layers.17.mlp.down_proj.scales": "model-00002-of-00003.safetensors",
|
| 283 |
+
"model.layers.17.mlp.gate_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 284 |
+
"model.layers.17.mlp.gate_proj.qweight": "model-00002-of-00003.safetensors",
|
| 285 |
+
"model.layers.17.mlp.gate_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 286 |
+
"model.layers.17.mlp.gate_proj.scales": "model-00002-of-00003.safetensors",
|
| 287 |
+
"model.layers.17.mlp.up_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 288 |
+
"model.layers.17.mlp.up_proj.qweight": "model-00002-of-00003.safetensors",
|
| 289 |
+
"model.layers.17.mlp.up_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 290 |
+
"model.layers.17.mlp.up_proj.scales": "model-00002-of-00003.safetensors",
|
| 291 |
+
"model.layers.17.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 292 |
+
"model.layers.17.self_attn.k_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 293 |
+
"model.layers.17.self_attn.k_proj.qweight": "model-00002-of-00003.safetensors",
|
| 294 |
+
"model.layers.17.self_attn.k_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 295 |
+
"model.layers.17.self_attn.k_proj.scales": "model-00002-of-00003.safetensors",
|
| 296 |
+
"model.layers.17.self_attn.o_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 297 |
+
"model.layers.17.self_attn.o_proj.qweight": "model-00002-of-00003.safetensors",
|
| 298 |
+
"model.layers.17.self_attn.o_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 299 |
+
"model.layers.17.self_attn.o_proj.scales": "model-00002-of-00003.safetensors",
|
| 300 |
+
"model.layers.17.self_attn.q_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 301 |
+
"model.layers.17.self_attn.q_proj.qweight": "model-00002-of-00003.safetensors",
|
| 302 |
+
"model.layers.17.self_attn.q_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 303 |
+
"model.layers.17.self_attn.q_proj.scales": "model-00002-of-00003.safetensors",
|
| 304 |
+
"model.layers.17.self_attn.v_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 305 |
+
"model.layers.17.self_attn.v_proj.qweight": "model-00002-of-00003.safetensors",
|
| 306 |
+
"model.layers.17.self_attn.v_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 307 |
+
"model.layers.17.self_attn.v_proj.scales": "model-00002-of-00003.safetensors",
|
| 308 |
+
"model.layers.18.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 309 |
+
"model.layers.18.mlp.down_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 310 |
+
"model.layers.18.mlp.down_proj.qweight": "model-00002-of-00003.safetensors",
|
| 311 |
+
"model.layers.18.mlp.down_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 312 |
+
"model.layers.18.mlp.down_proj.scales": "model-00002-of-00003.safetensors",
|
| 313 |
+
"model.layers.18.mlp.gate_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 314 |
+
"model.layers.18.mlp.gate_proj.qweight": "model-00002-of-00003.safetensors",
|
| 315 |
+
"model.layers.18.mlp.gate_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 316 |
+
"model.layers.18.mlp.gate_proj.scales": "model-00002-of-00003.safetensors",
|
| 317 |
+
"model.layers.18.mlp.up_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 318 |
+
"model.layers.18.mlp.up_proj.qweight": "model-00002-of-00003.safetensors",
|
| 319 |
+
"model.layers.18.mlp.up_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 320 |
+
"model.layers.18.mlp.up_proj.scales": "model-00002-of-00003.safetensors",
|
| 321 |
+
"model.layers.18.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 322 |
+
"model.layers.18.self_attn.k_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 323 |
+
"model.layers.18.self_attn.k_proj.qweight": "model-00002-of-00003.safetensors",
|
| 324 |
+
"model.layers.18.self_attn.k_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 325 |
+
"model.layers.18.self_attn.k_proj.scales": "model-00002-of-00003.safetensors",
|
| 326 |
+
"model.layers.18.self_attn.o_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 327 |
+
"model.layers.18.self_attn.o_proj.qweight": "model-00002-of-00003.safetensors",
|
| 328 |
+
"model.layers.18.self_attn.o_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 329 |
+
"model.layers.18.self_attn.o_proj.scales": "model-00002-of-00003.safetensors",
|
| 330 |
+
"model.layers.18.self_attn.q_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 331 |
+
"model.layers.18.self_attn.q_proj.qweight": "model-00002-of-00003.safetensors",
|
| 332 |
+
"model.layers.18.self_attn.q_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 333 |
+
"model.layers.18.self_attn.q_proj.scales": "model-00002-of-00003.safetensors",
|
| 334 |
+
"model.layers.18.self_attn.v_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 335 |
+
"model.layers.18.self_attn.v_proj.qweight": "model-00002-of-00003.safetensors",
|
| 336 |
+
"model.layers.18.self_attn.v_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 337 |
+
"model.layers.18.self_attn.v_proj.scales": "model-00002-of-00003.safetensors",
|
| 338 |
+
"model.layers.19.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 339 |
+
"model.layers.19.mlp.down_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 340 |
+
"model.layers.19.mlp.down_proj.qweight": "model-00002-of-00003.safetensors",
|
| 341 |
+
"model.layers.19.mlp.down_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 342 |
+
"model.layers.19.mlp.down_proj.scales": "model-00002-of-00003.safetensors",
|
| 343 |
+
"model.layers.19.mlp.gate_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 344 |
+
"model.layers.19.mlp.gate_proj.qweight": "model-00002-of-00003.safetensors",
|
| 345 |
+
"model.layers.19.mlp.gate_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 346 |
+
"model.layers.19.mlp.gate_proj.scales": "model-00002-of-00003.safetensors",
|
| 347 |
+
"model.layers.19.mlp.up_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 348 |
+
"model.layers.19.mlp.up_proj.qweight": "model-00002-of-00003.safetensors",
|
| 349 |
+
"model.layers.19.mlp.up_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 350 |
+
"model.layers.19.mlp.up_proj.scales": "model-00002-of-00003.safetensors",
|
| 351 |
+
"model.layers.19.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 352 |
+
"model.layers.19.self_attn.k_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 353 |
+
"model.layers.19.self_attn.k_proj.qweight": "model-00002-of-00003.safetensors",
|
| 354 |
+
"model.layers.19.self_attn.k_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 355 |
+
"model.layers.19.self_attn.k_proj.scales": "model-00002-of-00003.safetensors",
|
| 356 |
+
"model.layers.19.self_attn.o_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 357 |
+
"model.layers.19.self_attn.o_proj.qweight": "model-00002-of-00003.safetensors",
|
| 358 |
+
"model.layers.19.self_attn.o_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 359 |
+
"model.layers.19.self_attn.o_proj.scales": "model-00002-of-00003.safetensors",
|
| 360 |
+
"model.layers.19.self_attn.q_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 361 |
+
"model.layers.19.self_attn.q_proj.qweight": "model-00002-of-00003.safetensors",
|
| 362 |
+
"model.layers.19.self_attn.q_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 363 |
+
"model.layers.19.self_attn.q_proj.scales": "model-00002-of-00003.safetensors",
|
| 364 |
+
"model.layers.19.self_attn.v_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 365 |
+
"model.layers.19.self_attn.v_proj.qweight": "model-00002-of-00003.safetensors",
|
| 366 |
+
"model.layers.19.self_attn.v_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 367 |
+
"model.layers.19.self_attn.v_proj.scales": "model-00002-of-00003.safetensors",
|
| 368 |
+
"model.layers.2.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 369 |
+
"model.layers.2.mlp.down_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 370 |
+
"model.layers.2.mlp.down_proj.qweight": "model-00001-of-00003.safetensors",
|
| 371 |
+
"model.layers.2.mlp.down_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 372 |
+
"model.layers.2.mlp.down_proj.scales": "model-00001-of-00003.safetensors",
|
| 373 |
+
"model.layers.2.mlp.gate_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 374 |
+
"model.layers.2.mlp.gate_proj.qweight": "model-00001-of-00003.safetensors",
|
| 375 |
+
"model.layers.2.mlp.gate_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 376 |
+
"model.layers.2.mlp.gate_proj.scales": "model-00001-of-00003.safetensors",
|
| 377 |
+
"model.layers.2.mlp.up_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 378 |
+
"model.layers.2.mlp.up_proj.qweight": "model-00001-of-00003.safetensors",
|
| 379 |
+
"model.layers.2.mlp.up_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 380 |
+
"model.layers.2.mlp.up_proj.scales": "model-00001-of-00003.safetensors",
|
| 381 |
+
"model.layers.2.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 382 |
+
"model.layers.2.self_attn.k_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 383 |
+
"model.layers.2.self_attn.k_proj.qweight": "model-00001-of-00003.safetensors",
|
| 384 |
+
"model.layers.2.self_attn.k_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 385 |
+
"model.layers.2.self_attn.k_proj.scales": "model-00001-of-00003.safetensors",
|
| 386 |
+
"model.layers.2.self_attn.o_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 387 |
+
"model.layers.2.self_attn.o_proj.qweight": "model-00001-of-00003.safetensors",
|
| 388 |
+
"model.layers.2.self_attn.o_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 389 |
+
"model.layers.2.self_attn.o_proj.scales": "model-00001-of-00003.safetensors",
|
| 390 |
+
"model.layers.2.self_attn.q_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 391 |
+
"model.layers.2.self_attn.q_proj.qweight": "model-00001-of-00003.safetensors",
|
| 392 |
+
"model.layers.2.self_attn.q_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 393 |
+
"model.layers.2.self_attn.q_proj.scales": "model-00001-of-00003.safetensors",
|
| 394 |
+
"model.layers.2.self_attn.v_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 395 |
+
"model.layers.2.self_attn.v_proj.qweight": "model-00001-of-00003.safetensors",
|
| 396 |
+
"model.layers.2.self_attn.v_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 397 |
+
"model.layers.2.self_attn.v_proj.scales": "model-00001-of-00003.safetensors",
|
| 398 |
+
"model.layers.20.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 399 |
+
"model.layers.20.mlp.down_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 400 |
+
"model.layers.20.mlp.down_proj.qweight": "model-00002-of-00003.safetensors",
|
| 401 |
+
"model.layers.20.mlp.down_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 402 |
+
"model.layers.20.mlp.down_proj.scales": "model-00002-of-00003.safetensors",
|
| 403 |
+
"model.layers.20.mlp.gate_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 404 |
+
"model.layers.20.mlp.gate_proj.qweight": "model-00002-of-00003.safetensors",
|
| 405 |
+
"model.layers.20.mlp.gate_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 406 |
+
"model.layers.20.mlp.gate_proj.scales": "model-00002-of-00003.safetensors",
|
| 407 |
+
"model.layers.20.mlp.up_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 408 |
+
"model.layers.20.mlp.up_proj.qweight": "model-00002-of-00003.safetensors",
|
| 409 |
+
"model.layers.20.mlp.up_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 410 |
+
"model.layers.20.mlp.up_proj.scales": "model-00002-of-00003.safetensors",
|
| 411 |
+
"model.layers.20.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 412 |
+
"model.layers.20.self_attn.k_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 413 |
+
"model.layers.20.self_attn.k_proj.qweight": "model-00002-of-00003.safetensors",
|
| 414 |
+
"model.layers.20.self_attn.k_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 415 |
+
"model.layers.20.self_attn.k_proj.scales": "model-00002-of-00003.safetensors",
|
| 416 |
+
"model.layers.20.self_attn.o_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 417 |
+
"model.layers.20.self_attn.o_proj.qweight": "model-00002-of-00003.safetensors",
|
| 418 |
+
"model.layers.20.self_attn.o_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 419 |
+
"model.layers.20.self_attn.o_proj.scales": "model-00002-of-00003.safetensors",
|
| 420 |
+
"model.layers.20.self_attn.q_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 421 |
+
"model.layers.20.self_attn.q_proj.qweight": "model-00002-of-00003.safetensors",
|
| 422 |
+
"model.layers.20.self_attn.q_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 423 |
+
"model.layers.20.self_attn.q_proj.scales": "model-00002-of-00003.safetensors",
|
| 424 |
+
"model.layers.20.self_attn.v_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 425 |
+
"model.layers.20.self_attn.v_proj.qweight": "model-00002-of-00003.safetensors",
|
| 426 |
+
"model.layers.20.self_attn.v_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 427 |
+
"model.layers.20.self_attn.v_proj.scales": "model-00002-of-00003.safetensors",
|
| 428 |
+
"model.layers.21.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 429 |
+
"model.layers.21.mlp.down_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 430 |
+
"model.layers.21.mlp.down_proj.qweight": "model-00002-of-00003.safetensors",
|
| 431 |
+
"model.layers.21.mlp.down_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 432 |
+
"model.layers.21.mlp.down_proj.scales": "model-00002-of-00003.safetensors",
|
| 433 |
+
"model.layers.21.mlp.gate_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 434 |
+
"model.layers.21.mlp.gate_proj.qweight": "model-00002-of-00003.safetensors",
|
| 435 |
+
"model.layers.21.mlp.gate_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 436 |
+
"model.layers.21.mlp.gate_proj.scales": "model-00002-of-00003.safetensors",
|
| 437 |
+
"model.layers.21.mlp.up_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 438 |
+
"model.layers.21.mlp.up_proj.qweight": "model-00002-of-00003.safetensors",
|
| 439 |
+
"model.layers.21.mlp.up_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 440 |
+
"model.layers.21.mlp.up_proj.scales": "model-00002-of-00003.safetensors",
|
| 441 |
+
"model.layers.21.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 442 |
+
"model.layers.21.self_attn.k_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 443 |
+
"model.layers.21.self_attn.k_proj.qweight": "model-00002-of-00003.safetensors",
|
| 444 |
+
"model.layers.21.self_attn.k_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 445 |
+
"model.layers.21.self_attn.k_proj.scales": "model-00002-of-00003.safetensors",
|
| 446 |
+
"model.layers.21.self_attn.o_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 447 |
+
"model.layers.21.self_attn.o_proj.qweight": "model-00002-of-00003.safetensors",
|
| 448 |
+
"model.layers.21.self_attn.o_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 449 |
+
"model.layers.21.self_attn.o_proj.scales": "model-00002-of-00003.safetensors",
|
| 450 |
+
"model.layers.21.self_attn.q_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 451 |
+
"model.layers.21.self_attn.q_proj.qweight": "model-00002-of-00003.safetensors",
|
| 452 |
+
"model.layers.21.self_attn.q_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 453 |
+
"model.layers.21.self_attn.q_proj.scales": "model-00002-of-00003.safetensors",
|
| 454 |
+
"model.layers.21.self_attn.v_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 455 |
+
"model.layers.21.self_attn.v_proj.qweight": "model-00002-of-00003.safetensors",
|
| 456 |
+
"model.layers.21.self_attn.v_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 457 |
+
"model.layers.21.self_attn.v_proj.scales": "model-00002-of-00003.safetensors",
|
| 458 |
+
"model.layers.22.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 459 |
+
"model.layers.22.mlp.down_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 460 |
+
"model.layers.22.mlp.down_proj.qweight": "model-00002-of-00003.safetensors",
|
| 461 |
+
"model.layers.22.mlp.down_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 462 |
+
"model.layers.22.mlp.down_proj.scales": "model-00002-of-00003.safetensors",
|
| 463 |
+
"model.layers.22.mlp.gate_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 464 |
+
"model.layers.22.mlp.gate_proj.qweight": "model-00002-of-00003.safetensors",
|
| 465 |
+
"model.layers.22.mlp.gate_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 466 |
+
"model.layers.22.mlp.gate_proj.scales": "model-00002-of-00003.safetensors",
|
| 467 |
+
"model.layers.22.mlp.up_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 468 |
+
"model.layers.22.mlp.up_proj.qweight": "model-00002-of-00003.safetensors",
|
| 469 |
+
"model.layers.22.mlp.up_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 470 |
+
"model.layers.22.mlp.up_proj.scales": "model-00002-of-00003.safetensors",
|
| 471 |
+
"model.layers.22.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 472 |
+
"model.layers.22.self_attn.k_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 473 |
+
"model.layers.22.self_attn.k_proj.qweight": "model-00002-of-00003.safetensors",
|
| 474 |
+
"model.layers.22.self_attn.k_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 475 |
+
"model.layers.22.self_attn.k_proj.scales": "model-00002-of-00003.safetensors",
|
| 476 |
+
"model.layers.22.self_attn.o_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 477 |
+
"model.layers.22.self_attn.o_proj.qweight": "model-00002-of-00003.safetensors",
|
| 478 |
+
"model.layers.22.self_attn.o_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 479 |
+
"model.layers.22.self_attn.o_proj.scales": "model-00002-of-00003.safetensors",
|
| 480 |
+
"model.layers.22.self_attn.q_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 481 |
+
"model.layers.22.self_attn.q_proj.qweight": "model-00002-of-00003.safetensors",
|
| 482 |
+
"model.layers.22.self_attn.q_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 483 |
+
"model.layers.22.self_attn.q_proj.scales": "model-00002-of-00003.safetensors",
|
| 484 |
+
"model.layers.22.self_attn.v_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 485 |
+
"model.layers.22.self_attn.v_proj.qweight": "model-00002-of-00003.safetensors",
|
| 486 |
+
"model.layers.22.self_attn.v_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 487 |
+
"model.layers.22.self_attn.v_proj.scales": "model-00002-of-00003.safetensors",
|
| 488 |
+
"model.layers.23.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 489 |
+
"model.layers.23.mlp.down_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 490 |
+
"model.layers.23.mlp.down_proj.qweight": "model-00002-of-00003.safetensors",
|
| 491 |
+
"model.layers.23.mlp.down_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 492 |
+
"model.layers.23.mlp.down_proj.scales": "model-00002-of-00003.safetensors",
|
| 493 |
+
"model.layers.23.mlp.gate_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 494 |
+
"model.layers.23.mlp.gate_proj.qweight": "model-00002-of-00003.safetensors",
|
| 495 |
+
"model.layers.23.mlp.gate_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 496 |
+
"model.layers.23.mlp.gate_proj.scales": "model-00002-of-00003.safetensors",
|
| 497 |
+
"model.layers.23.mlp.up_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 498 |
+
"model.layers.23.mlp.up_proj.qweight": "model-00002-of-00003.safetensors",
|
| 499 |
+
"model.layers.23.mlp.up_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 500 |
+
"model.layers.23.mlp.up_proj.scales": "model-00002-of-00003.safetensors",
|
| 501 |
+
"model.layers.23.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 502 |
+
"model.layers.23.self_attn.k_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 503 |
+
"model.layers.23.self_attn.k_proj.qweight": "model-00002-of-00003.safetensors",
|
| 504 |
+
"model.layers.23.self_attn.k_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 505 |
+
"model.layers.23.self_attn.k_proj.scales": "model-00002-of-00003.safetensors",
|
| 506 |
+
"model.layers.23.self_attn.o_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 507 |
+
"model.layers.23.self_attn.o_proj.qweight": "model-00002-of-00003.safetensors",
|
| 508 |
+
"model.layers.23.self_attn.o_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 509 |
+
"model.layers.23.self_attn.o_proj.scales": "model-00002-of-00003.safetensors",
|
| 510 |
+
"model.layers.23.self_attn.q_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 511 |
+
"model.layers.23.self_attn.q_proj.qweight": "model-00002-of-00003.safetensors",
|
| 512 |
+
"model.layers.23.self_attn.q_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 513 |
+
"model.layers.23.self_attn.q_proj.scales": "model-00002-of-00003.safetensors",
|
| 514 |
+
"model.layers.23.self_attn.v_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 515 |
+
"model.layers.23.self_attn.v_proj.qweight": "model-00002-of-00003.safetensors",
|
| 516 |
+
"model.layers.23.self_attn.v_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 517 |
+
"model.layers.23.self_attn.v_proj.scales": "model-00002-of-00003.safetensors",
|
| 518 |
+
"model.layers.24.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 519 |
+
"model.layers.24.mlp.down_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 520 |
+
"model.layers.24.mlp.down_proj.qweight": "model-00002-of-00003.safetensors",
|
| 521 |
+
"model.layers.24.mlp.down_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 522 |
+
"model.layers.24.mlp.down_proj.scales": "model-00002-of-00003.safetensors",
|
| 523 |
+
"model.layers.24.mlp.gate_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 524 |
+
"model.layers.24.mlp.gate_proj.qweight": "model-00002-of-00003.safetensors",
|
| 525 |
+
"model.layers.24.mlp.gate_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 526 |
+
"model.layers.24.mlp.gate_proj.scales": "model-00002-of-00003.safetensors",
|
| 527 |
+
"model.layers.24.mlp.up_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 528 |
+
"model.layers.24.mlp.up_proj.qweight": "model-00002-of-00003.safetensors",
|
| 529 |
+
"model.layers.24.mlp.up_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 530 |
+
"model.layers.24.mlp.up_proj.scales": "model-00002-of-00003.safetensors",
|
| 531 |
+
"model.layers.24.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 532 |
+
"model.layers.24.self_attn.k_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 533 |
+
"model.layers.24.self_attn.k_proj.qweight": "model-00002-of-00003.safetensors",
|
| 534 |
+
"model.layers.24.self_attn.k_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 535 |
+
"model.layers.24.self_attn.k_proj.scales": "model-00002-of-00003.safetensors",
|
| 536 |
+
"model.layers.24.self_attn.o_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 537 |
+
"model.layers.24.self_attn.o_proj.qweight": "model-00002-of-00003.safetensors",
|
| 538 |
+
"model.layers.24.self_attn.o_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 539 |
+
"model.layers.24.self_attn.o_proj.scales": "model-00002-of-00003.safetensors",
|
| 540 |
+
"model.layers.24.self_attn.q_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 541 |
+
"model.layers.24.self_attn.q_proj.qweight": "model-00002-of-00003.safetensors",
|
| 542 |
+
"model.layers.24.self_attn.q_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 543 |
+
"model.layers.24.self_attn.q_proj.scales": "model-00002-of-00003.safetensors",
|
| 544 |
+
"model.layers.24.self_attn.v_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 545 |
+
"model.layers.24.self_attn.v_proj.qweight": "model-00002-of-00003.safetensors",
|
| 546 |
+
"model.layers.24.self_attn.v_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 547 |
+
"model.layers.24.self_attn.v_proj.scales": "model-00002-of-00003.safetensors",
|
| 548 |
+
"model.layers.25.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 549 |
+
"model.layers.25.mlp.down_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 550 |
+
"model.layers.25.mlp.down_proj.qweight": "model-00002-of-00003.safetensors",
|
| 551 |
+
"model.layers.25.mlp.down_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 552 |
+
"model.layers.25.mlp.down_proj.scales": "model-00002-of-00003.safetensors",
|
| 553 |
+
"model.layers.25.mlp.gate_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 554 |
+
"model.layers.25.mlp.gate_proj.qweight": "model-00002-of-00003.safetensors",
|
| 555 |
+
"model.layers.25.mlp.gate_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 556 |
+
"model.layers.25.mlp.gate_proj.scales": "model-00002-of-00003.safetensors",
|
| 557 |
+
"model.layers.25.mlp.up_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 558 |
+
"model.layers.25.mlp.up_proj.qweight": "model-00002-of-00003.safetensors",
|
| 559 |
+
"model.layers.25.mlp.up_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 560 |
+
"model.layers.25.mlp.up_proj.scales": "model-00002-of-00003.safetensors",
|
| 561 |
+
"model.layers.25.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 562 |
+
"model.layers.25.self_attn.k_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 563 |
+
"model.layers.25.self_attn.k_proj.qweight": "model-00002-of-00003.safetensors",
|
| 564 |
+
"model.layers.25.self_attn.k_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 565 |
+
"model.layers.25.self_attn.k_proj.scales": "model-00002-of-00003.safetensors",
|
| 566 |
+
"model.layers.25.self_attn.o_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 567 |
+
"model.layers.25.self_attn.o_proj.qweight": "model-00002-of-00003.safetensors",
|
| 568 |
+
"model.layers.25.self_attn.o_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 569 |
+
"model.layers.25.self_attn.o_proj.scales": "model-00002-of-00003.safetensors",
|
| 570 |
+
"model.layers.25.self_attn.q_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 571 |
+
"model.layers.25.self_attn.q_proj.qweight": "model-00002-of-00003.safetensors",
|
| 572 |
+
"model.layers.25.self_attn.q_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 573 |
+
"model.layers.25.self_attn.q_proj.scales": "model-00002-of-00003.safetensors",
|
| 574 |
+
"model.layers.25.self_attn.v_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 575 |
+
"model.layers.25.self_attn.v_proj.qweight": "model-00002-of-00003.safetensors",
|
| 576 |
+
"model.layers.25.self_attn.v_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 577 |
+
"model.layers.25.self_attn.v_proj.scales": "model-00002-of-00003.safetensors",
|
| 578 |
+
"model.layers.26.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 579 |
+
"model.layers.26.mlp.down_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 580 |
+
"model.layers.26.mlp.down_proj.qweight": "model-00002-of-00003.safetensors",
|
| 581 |
+
"model.layers.26.mlp.down_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 582 |
+
"model.layers.26.mlp.down_proj.scales": "model-00002-of-00003.safetensors",
|
| 583 |
+
"model.layers.26.mlp.gate_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 584 |
+
"model.layers.26.mlp.gate_proj.qweight": "model-00002-of-00003.safetensors",
|
| 585 |
+
"model.layers.26.mlp.gate_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 586 |
+
"model.layers.26.mlp.gate_proj.scales": "model-00002-of-00003.safetensors",
|
| 587 |
+
"model.layers.26.mlp.up_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 588 |
+
"model.layers.26.mlp.up_proj.qweight": "model-00002-of-00003.safetensors",
|
| 589 |
+
"model.layers.26.mlp.up_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 590 |
+
"model.layers.26.mlp.up_proj.scales": "model-00002-of-00003.safetensors",
|
| 591 |
+
"model.layers.26.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 592 |
+
"model.layers.26.self_attn.k_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 593 |
+
"model.layers.26.self_attn.k_proj.qweight": "model-00002-of-00003.safetensors",
|
| 594 |
+
"model.layers.26.self_attn.k_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 595 |
+
"model.layers.26.self_attn.k_proj.scales": "model-00002-of-00003.safetensors",
|
| 596 |
+
"model.layers.26.self_attn.o_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 597 |
+
"model.layers.26.self_attn.o_proj.qweight": "model-00002-of-00003.safetensors",
|
| 598 |
+
"model.layers.26.self_attn.o_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 599 |
+
"model.layers.26.self_attn.o_proj.scales": "model-00002-of-00003.safetensors",
|
| 600 |
+
"model.layers.26.self_attn.q_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 601 |
+
"model.layers.26.self_attn.q_proj.qweight": "model-00002-of-00003.safetensors",
|
| 602 |
+
"model.layers.26.self_attn.q_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 603 |
+
"model.layers.26.self_attn.q_proj.scales": "model-00002-of-00003.safetensors",
|
| 604 |
+
"model.layers.26.self_attn.v_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 605 |
+
"model.layers.26.self_attn.v_proj.qweight": "model-00002-of-00003.safetensors",
|
| 606 |
+
"model.layers.26.self_attn.v_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 607 |
+
"model.layers.26.self_attn.v_proj.scales": "model-00002-of-00003.safetensors",
|
| 608 |
+
"model.layers.27.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 609 |
+
"model.layers.27.mlp.down_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 610 |
+
"model.layers.27.mlp.down_proj.qweight": "model-00002-of-00003.safetensors",
|
| 611 |
+
"model.layers.27.mlp.down_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 612 |
+
"model.layers.27.mlp.down_proj.scales": "model-00002-of-00003.safetensors",
|
| 613 |
+
"model.layers.27.mlp.gate_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 614 |
+
"model.layers.27.mlp.gate_proj.qweight": "model-00002-of-00003.safetensors",
|
| 615 |
+
"model.layers.27.mlp.gate_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 616 |
+
"model.layers.27.mlp.gate_proj.scales": "model-00002-of-00003.safetensors",
|
| 617 |
+
"model.layers.27.mlp.up_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 618 |
+
"model.layers.27.mlp.up_proj.qweight": "model-00002-of-00003.safetensors",
|
| 619 |
+
"model.layers.27.mlp.up_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 620 |
+
"model.layers.27.mlp.up_proj.scales": "model-00002-of-00003.safetensors",
|
| 621 |
+
"model.layers.27.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 622 |
+
"model.layers.27.self_attn.k_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 623 |
+
"model.layers.27.self_attn.k_proj.qweight": "model-00002-of-00003.safetensors",
|
| 624 |
+
"model.layers.27.self_attn.k_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 625 |
+
"model.layers.27.self_attn.k_proj.scales": "model-00002-of-00003.safetensors",
|
| 626 |
+
"model.layers.27.self_attn.o_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 627 |
+
"model.layers.27.self_attn.o_proj.qweight": "model-00002-of-00003.safetensors",
|
| 628 |
+
"model.layers.27.self_attn.o_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 629 |
+
"model.layers.27.self_attn.o_proj.scales": "model-00002-of-00003.safetensors",
|
| 630 |
+
"model.layers.27.self_attn.q_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 631 |
+
"model.layers.27.self_attn.q_proj.qweight": "model-00002-of-00003.safetensors",
|
| 632 |
+
"model.layers.27.self_attn.q_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 633 |
+
"model.layers.27.self_attn.q_proj.scales": "model-00002-of-00003.safetensors",
|
| 634 |
+
"model.layers.27.self_attn.v_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 635 |
+
"model.layers.27.self_attn.v_proj.qweight": "model-00002-of-00003.safetensors",
|
| 636 |
+
"model.layers.27.self_attn.v_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 637 |
+
"model.layers.27.self_attn.v_proj.scales": "model-00002-of-00003.safetensors",
|
| 638 |
+
"model.layers.28.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 639 |
+
"model.layers.28.mlp.down_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 640 |
+
"model.layers.28.mlp.down_proj.qweight": "model-00002-of-00003.safetensors",
|
| 641 |
+
"model.layers.28.mlp.down_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 642 |
+
"model.layers.28.mlp.down_proj.scales": "model-00002-of-00003.safetensors",
|
| 643 |
+
"model.layers.28.mlp.gate_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 644 |
+
"model.layers.28.mlp.gate_proj.qweight": "model-00002-of-00003.safetensors",
|
| 645 |
+
"model.layers.28.mlp.gate_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 646 |
+
"model.layers.28.mlp.gate_proj.scales": "model-00002-of-00003.safetensors",
|
| 647 |
+
"model.layers.28.mlp.up_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 648 |
+
"model.layers.28.mlp.up_proj.qweight": "model-00002-of-00003.safetensors",
|
| 649 |
+
"model.layers.28.mlp.up_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 650 |
+
"model.layers.28.mlp.up_proj.scales": "model-00002-of-00003.safetensors",
|
| 651 |
+
"model.layers.28.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 652 |
+
"model.layers.28.self_attn.k_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 653 |
+
"model.layers.28.self_attn.k_proj.qweight": "model-00002-of-00003.safetensors",
|
| 654 |
+
"model.layers.28.self_attn.k_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 655 |
+
"model.layers.28.self_attn.k_proj.scales": "model-00002-of-00003.safetensors",
|
| 656 |
+
"model.layers.28.self_attn.o_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 657 |
+
"model.layers.28.self_attn.o_proj.qweight": "model-00002-of-00003.safetensors",
|
| 658 |
+
"model.layers.28.self_attn.o_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 659 |
+
"model.layers.28.self_attn.o_proj.scales": "model-00002-of-00003.safetensors",
|
| 660 |
+
"model.layers.28.self_attn.q_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 661 |
+
"model.layers.28.self_attn.q_proj.qweight": "model-00002-of-00003.safetensors",
|
| 662 |
+
"model.layers.28.self_attn.q_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 663 |
+
"model.layers.28.self_attn.q_proj.scales": "model-00002-of-00003.safetensors",
|
| 664 |
+
"model.layers.28.self_attn.v_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 665 |
+
"model.layers.28.self_attn.v_proj.qweight": "model-00002-of-00003.safetensors",
|
| 666 |
+
"model.layers.28.self_attn.v_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 667 |
+
"model.layers.28.self_attn.v_proj.scales": "model-00002-of-00003.safetensors",
|
| 668 |
+
"model.layers.29.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 669 |
+
"model.layers.29.mlp.down_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 670 |
+
"model.layers.29.mlp.down_proj.qweight": "model-00002-of-00003.safetensors",
|
| 671 |
+
"model.layers.29.mlp.down_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 672 |
+
"model.layers.29.mlp.down_proj.scales": "model-00002-of-00003.safetensors",
|
| 673 |
+
"model.layers.29.mlp.gate_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 674 |
+
"model.layers.29.mlp.gate_proj.qweight": "model-00002-of-00003.safetensors",
|
| 675 |
+
"model.layers.29.mlp.gate_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 676 |
+
"model.layers.29.mlp.gate_proj.scales": "model-00002-of-00003.safetensors",
|
| 677 |
+
"model.layers.29.mlp.up_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 678 |
+
"model.layers.29.mlp.up_proj.qweight": "model-00002-of-00003.safetensors",
|
| 679 |
+
"model.layers.29.mlp.up_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 680 |
+
"model.layers.29.mlp.up_proj.scales": "model-00002-of-00003.safetensors",
|
| 681 |
+
"model.layers.29.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 682 |
+
"model.layers.29.self_attn.k_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 683 |
+
"model.layers.29.self_attn.k_proj.qweight": "model-00002-of-00003.safetensors",
|
| 684 |
+
"model.layers.29.self_attn.k_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 685 |
+
"model.layers.29.self_attn.k_proj.scales": "model-00002-of-00003.safetensors",
|
| 686 |
+
"model.layers.29.self_attn.o_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 687 |
+
"model.layers.29.self_attn.o_proj.qweight": "model-00002-of-00003.safetensors",
|
| 688 |
+
"model.layers.29.self_attn.o_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 689 |
+
"model.layers.29.self_attn.o_proj.scales": "model-00002-of-00003.safetensors",
|
| 690 |
+
"model.layers.29.self_attn.q_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 691 |
+
"model.layers.29.self_attn.q_proj.qweight": "model-00002-of-00003.safetensors",
|
| 692 |
+
"model.layers.29.self_attn.q_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 693 |
+
"model.layers.29.self_attn.q_proj.scales": "model-00002-of-00003.safetensors",
|
| 694 |
+
"model.layers.29.self_attn.v_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 695 |
+
"model.layers.29.self_attn.v_proj.qweight": "model-00002-of-00003.safetensors",
|
| 696 |
+
"model.layers.29.self_attn.v_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 697 |
+
"model.layers.29.self_attn.v_proj.scales": "model-00002-of-00003.safetensors",
|
| 698 |
+
"model.layers.3.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 699 |
+
"model.layers.3.mlp.down_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 700 |
+
"model.layers.3.mlp.down_proj.qweight": "model-00001-of-00003.safetensors",
|
| 701 |
+
"model.layers.3.mlp.down_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 702 |
+
"model.layers.3.mlp.down_proj.scales": "model-00001-of-00003.safetensors",
|
| 703 |
+
"model.layers.3.mlp.gate_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 704 |
+
"model.layers.3.mlp.gate_proj.qweight": "model-00001-of-00003.safetensors",
|
| 705 |
+
"model.layers.3.mlp.gate_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 706 |
+
"model.layers.3.mlp.gate_proj.scales": "model-00001-of-00003.safetensors",
|
| 707 |
+
"model.layers.3.mlp.up_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 708 |
+
"model.layers.3.mlp.up_proj.qweight": "model-00001-of-00003.safetensors",
|
| 709 |
+
"model.layers.3.mlp.up_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 710 |
+
"model.layers.3.mlp.up_proj.scales": "model-00001-of-00003.safetensors",
|
| 711 |
+
"model.layers.3.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 712 |
+
"model.layers.3.self_attn.k_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 713 |
+
"model.layers.3.self_attn.k_proj.qweight": "model-00001-of-00003.safetensors",
|
| 714 |
+
"model.layers.3.self_attn.k_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 715 |
+
"model.layers.3.self_attn.k_proj.scales": "model-00001-of-00003.safetensors",
|
| 716 |
+
"model.layers.3.self_attn.o_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 717 |
+
"model.layers.3.self_attn.o_proj.qweight": "model-00001-of-00003.safetensors",
|
| 718 |
+
"model.layers.3.self_attn.o_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 719 |
+
"model.layers.3.self_attn.o_proj.scales": "model-00001-of-00003.safetensors",
|
| 720 |
+
"model.layers.3.self_attn.q_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 721 |
+
"model.layers.3.self_attn.q_proj.qweight": "model-00001-of-00003.safetensors",
|
| 722 |
+
"model.layers.3.self_attn.q_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 723 |
+
"model.layers.3.self_attn.q_proj.scales": "model-00001-of-00003.safetensors",
|
| 724 |
+
"model.layers.3.self_attn.v_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 725 |
+
"model.layers.3.self_attn.v_proj.qweight": "model-00001-of-00003.safetensors",
|
| 726 |
+
"model.layers.3.self_attn.v_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 727 |
+
"model.layers.3.self_attn.v_proj.scales": "model-00001-of-00003.safetensors",
|
| 728 |
+
"model.layers.30.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 729 |
+
"model.layers.30.mlp.down_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 730 |
+
"model.layers.30.mlp.down_proj.qweight": "model-00002-of-00003.safetensors",
|
| 731 |
+
"model.layers.30.mlp.down_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 732 |
+
"model.layers.30.mlp.down_proj.scales": "model-00002-of-00003.safetensors",
|
| 733 |
+
"model.layers.30.mlp.gate_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 734 |
+
"model.layers.30.mlp.gate_proj.qweight": "model-00002-of-00003.safetensors",
|
| 735 |
+
"model.layers.30.mlp.gate_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 736 |
+
"model.layers.30.mlp.gate_proj.scales": "model-00002-of-00003.safetensors",
|
| 737 |
+
"model.layers.30.mlp.up_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 738 |
+
"model.layers.30.mlp.up_proj.qweight": "model-00002-of-00003.safetensors",
|
| 739 |
+
"model.layers.30.mlp.up_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 740 |
+
"model.layers.30.mlp.up_proj.scales": "model-00002-of-00003.safetensors",
|
| 741 |
+
"model.layers.30.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
| 742 |
+
"model.layers.30.self_attn.k_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 743 |
+
"model.layers.30.self_attn.k_proj.qweight": "model-00002-of-00003.safetensors",
|
| 744 |
+
"model.layers.30.self_attn.k_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 745 |
+
"model.layers.30.self_attn.k_proj.scales": "model-00002-of-00003.safetensors",
|
| 746 |
+
"model.layers.30.self_attn.o_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 747 |
+
"model.layers.30.self_attn.o_proj.qweight": "model-00002-of-00003.safetensors",
|
| 748 |
+
"model.layers.30.self_attn.o_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 749 |
+
"model.layers.30.self_attn.o_proj.scales": "model-00002-of-00003.safetensors",
|
| 750 |
+
"model.layers.30.self_attn.q_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 751 |
+
"model.layers.30.self_attn.q_proj.qweight": "model-00002-of-00003.safetensors",
|
| 752 |
+
"model.layers.30.self_attn.q_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 753 |
+
"model.layers.30.self_attn.q_proj.scales": "model-00002-of-00003.safetensors",
|
| 754 |
+
"model.layers.30.self_attn.v_proj.g_idx": "model-00002-of-00003.safetensors",
|
| 755 |
+
"model.layers.30.self_attn.v_proj.qweight": "model-00002-of-00003.safetensors",
|
| 756 |
+
"model.layers.30.self_attn.v_proj.qzeros": "model-00002-of-00003.safetensors",
|
| 757 |
+
"model.layers.30.self_attn.v_proj.scales": "model-00002-of-00003.safetensors",
|
| 758 |
+
"model.layers.31.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 759 |
+
"model.layers.31.mlp.down_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 760 |
+
"model.layers.31.mlp.down_proj.qweight": "model-00003-of-00003.safetensors",
|
| 761 |
+
"model.layers.31.mlp.down_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 762 |
+
"model.layers.31.mlp.down_proj.scales": "model-00003-of-00003.safetensors",
|
| 763 |
+
"model.layers.31.mlp.gate_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 764 |
+
"model.layers.31.mlp.gate_proj.qweight": "model-00003-of-00003.safetensors",
|
| 765 |
+
"model.layers.31.mlp.gate_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 766 |
+
"model.layers.31.mlp.gate_proj.scales": "model-00003-of-00003.safetensors",
|
| 767 |
+
"model.layers.31.mlp.up_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 768 |
+
"model.layers.31.mlp.up_proj.qweight": "model-00003-of-00003.safetensors",
|
| 769 |
+
"model.layers.31.mlp.up_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 770 |
+
"model.layers.31.mlp.up_proj.scales": "model-00003-of-00003.safetensors",
|
| 771 |
+
"model.layers.31.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 772 |
+
"model.layers.31.self_attn.k_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 773 |
+
"model.layers.31.self_attn.k_proj.qweight": "model-00003-of-00003.safetensors",
|
| 774 |
+
"model.layers.31.self_attn.k_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 775 |
+
"model.layers.31.self_attn.k_proj.scales": "model-00003-of-00003.safetensors",
|
| 776 |
+
"model.layers.31.self_attn.o_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 777 |
+
"model.layers.31.self_attn.o_proj.qweight": "model-00003-of-00003.safetensors",
|
| 778 |
+
"model.layers.31.self_attn.o_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 779 |
+
"model.layers.31.self_attn.o_proj.scales": "model-00003-of-00003.safetensors",
|
| 780 |
+
"model.layers.31.self_attn.q_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 781 |
+
"model.layers.31.self_attn.q_proj.qweight": "model-00003-of-00003.safetensors",
|
| 782 |
+
"model.layers.31.self_attn.q_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 783 |
+
"model.layers.31.self_attn.q_proj.scales": "model-00003-of-00003.safetensors",
|
| 784 |
+
"model.layers.31.self_attn.v_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 785 |
+
"model.layers.31.self_attn.v_proj.qweight": "model-00003-of-00003.safetensors",
|
| 786 |
+
"model.layers.31.self_attn.v_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 787 |
+
"model.layers.31.self_attn.v_proj.scales": "model-00003-of-00003.safetensors",
|
| 788 |
+
"model.layers.32.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 789 |
+
"model.layers.32.mlp.down_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 790 |
+
"model.layers.32.mlp.down_proj.qweight": "model-00003-of-00003.safetensors",
|
| 791 |
+
"model.layers.32.mlp.down_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 792 |
+
"model.layers.32.mlp.down_proj.scales": "model-00003-of-00003.safetensors",
|
| 793 |
+
"model.layers.32.mlp.gate_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 794 |
+
"model.layers.32.mlp.gate_proj.qweight": "model-00003-of-00003.safetensors",
|
| 795 |
+
"model.layers.32.mlp.gate_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 796 |
+
"model.layers.32.mlp.gate_proj.scales": "model-00003-of-00003.safetensors",
|
| 797 |
+
"model.layers.32.mlp.up_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 798 |
+
"model.layers.32.mlp.up_proj.qweight": "model-00003-of-00003.safetensors",
|
| 799 |
+
"model.layers.32.mlp.up_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 800 |
+
"model.layers.32.mlp.up_proj.scales": "model-00003-of-00003.safetensors",
|
| 801 |
+
"model.layers.32.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 802 |
+
"model.layers.32.self_attn.k_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 803 |
+
"model.layers.32.self_attn.k_proj.qweight": "model-00003-of-00003.safetensors",
|
| 804 |
+
"model.layers.32.self_attn.k_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 805 |
+
"model.layers.32.self_attn.k_proj.scales": "model-00003-of-00003.safetensors",
|
| 806 |
+
"model.layers.32.self_attn.o_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 807 |
+
"model.layers.32.self_attn.o_proj.qweight": "model-00003-of-00003.safetensors",
|
| 808 |
+
"model.layers.32.self_attn.o_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 809 |
+
"model.layers.32.self_attn.o_proj.scales": "model-00003-of-00003.safetensors",
|
| 810 |
+
"model.layers.32.self_attn.q_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 811 |
+
"model.layers.32.self_attn.q_proj.qweight": "model-00003-of-00003.safetensors",
|
| 812 |
+
"model.layers.32.self_attn.q_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 813 |
+
"model.layers.32.self_attn.q_proj.scales": "model-00003-of-00003.safetensors",
|
| 814 |
+
"model.layers.32.self_attn.v_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 815 |
+
"model.layers.32.self_attn.v_proj.qweight": "model-00003-of-00003.safetensors",
|
| 816 |
+
"model.layers.32.self_attn.v_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 817 |
+
"model.layers.32.self_attn.v_proj.scales": "model-00003-of-00003.safetensors",
|
| 818 |
+
"model.layers.33.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 819 |
+
"model.layers.33.mlp.down_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 820 |
+
"model.layers.33.mlp.down_proj.qweight": "model-00003-of-00003.safetensors",
|
| 821 |
+
"model.layers.33.mlp.down_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 822 |
+
"model.layers.33.mlp.down_proj.scales": "model-00003-of-00003.safetensors",
|
| 823 |
+
"model.layers.33.mlp.gate_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 824 |
+
"model.layers.33.mlp.gate_proj.qweight": "model-00003-of-00003.safetensors",
|
| 825 |
+
"model.layers.33.mlp.gate_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 826 |
+
"model.layers.33.mlp.gate_proj.scales": "model-00003-of-00003.safetensors",
|
| 827 |
+
"model.layers.33.mlp.up_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 828 |
+
"model.layers.33.mlp.up_proj.qweight": "model-00003-of-00003.safetensors",
|
| 829 |
+
"model.layers.33.mlp.up_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 830 |
+
"model.layers.33.mlp.up_proj.scales": "model-00003-of-00003.safetensors",
|
| 831 |
+
"model.layers.33.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 832 |
+
"model.layers.33.self_attn.k_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 833 |
+
"model.layers.33.self_attn.k_proj.qweight": "model-00003-of-00003.safetensors",
|
| 834 |
+
"model.layers.33.self_attn.k_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 835 |
+
"model.layers.33.self_attn.k_proj.scales": "model-00003-of-00003.safetensors",
|
| 836 |
+
"model.layers.33.self_attn.o_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 837 |
+
"model.layers.33.self_attn.o_proj.qweight": "model-00003-of-00003.safetensors",
|
| 838 |
+
"model.layers.33.self_attn.o_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 839 |
+
"model.layers.33.self_attn.o_proj.scales": "model-00003-of-00003.safetensors",
|
| 840 |
+
"model.layers.33.self_attn.q_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 841 |
+
"model.layers.33.self_attn.q_proj.qweight": "model-00003-of-00003.safetensors",
|
| 842 |
+
"model.layers.33.self_attn.q_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 843 |
+
"model.layers.33.self_attn.q_proj.scales": "model-00003-of-00003.safetensors",
|
| 844 |
+
"model.layers.33.self_attn.v_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 845 |
+
"model.layers.33.self_attn.v_proj.qweight": "model-00003-of-00003.safetensors",
|
| 846 |
+
"model.layers.33.self_attn.v_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 847 |
+
"model.layers.33.self_attn.v_proj.scales": "model-00003-of-00003.safetensors",
|
| 848 |
+
"model.layers.34.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 849 |
+
"model.layers.34.mlp.down_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 850 |
+
"model.layers.34.mlp.down_proj.qweight": "model-00003-of-00003.safetensors",
|
| 851 |
+
"model.layers.34.mlp.down_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 852 |
+
"model.layers.34.mlp.down_proj.scales": "model-00003-of-00003.safetensors",
|
| 853 |
+
"model.layers.34.mlp.gate_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 854 |
+
"model.layers.34.mlp.gate_proj.qweight": "model-00003-of-00003.safetensors",
|
| 855 |
+
"model.layers.34.mlp.gate_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 856 |
+
"model.layers.34.mlp.gate_proj.scales": "model-00003-of-00003.safetensors",
|
| 857 |
+
"model.layers.34.mlp.up_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 858 |
+
"model.layers.34.mlp.up_proj.qweight": "model-00003-of-00003.safetensors",
|
| 859 |
+
"model.layers.34.mlp.up_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 860 |
+
"model.layers.34.mlp.up_proj.scales": "model-00003-of-00003.safetensors",
|
| 861 |
+
"model.layers.34.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 862 |
+
"model.layers.34.self_attn.k_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 863 |
+
"model.layers.34.self_attn.k_proj.qweight": "model-00003-of-00003.safetensors",
|
| 864 |
+
"model.layers.34.self_attn.k_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 865 |
+
"model.layers.34.self_attn.k_proj.scales": "model-00003-of-00003.safetensors",
|
| 866 |
+
"model.layers.34.self_attn.o_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 867 |
+
"model.layers.34.self_attn.o_proj.qweight": "model-00003-of-00003.safetensors",
|
| 868 |
+
"model.layers.34.self_attn.o_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 869 |
+
"model.layers.34.self_attn.o_proj.scales": "model-00003-of-00003.safetensors",
|
| 870 |
+
"model.layers.34.self_attn.q_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 871 |
+
"model.layers.34.self_attn.q_proj.qweight": "model-00003-of-00003.safetensors",
|
| 872 |
+
"model.layers.34.self_attn.q_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 873 |
+
"model.layers.34.self_attn.q_proj.scales": "model-00003-of-00003.safetensors",
|
| 874 |
+
"model.layers.34.self_attn.v_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 875 |
+
"model.layers.34.self_attn.v_proj.qweight": "model-00003-of-00003.safetensors",
|
| 876 |
+
"model.layers.34.self_attn.v_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 877 |
+
"model.layers.34.self_attn.v_proj.scales": "model-00003-of-00003.safetensors",
|
| 878 |
+
"model.layers.35.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 879 |
+
"model.layers.35.mlp.down_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 880 |
+
"model.layers.35.mlp.down_proj.qweight": "model-00003-of-00003.safetensors",
|
| 881 |
+
"model.layers.35.mlp.down_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 882 |
+
"model.layers.35.mlp.down_proj.scales": "model-00003-of-00003.safetensors",
|
| 883 |
+
"model.layers.35.mlp.gate_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 884 |
+
"model.layers.35.mlp.gate_proj.qweight": "model-00003-of-00003.safetensors",
|
| 885 |
+
"model.layers.35.mlp.gate_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 886 |
+
"model.layers.35.mlp.gate_proj.scales": "model-00003-of-00003.safetensors",
|
| 887 |
+
"model.layers.35.mlp.up_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 888 |
+
"model.layers.35.mlp.up_proj.qweight": "model-00003-of-00003.safetensors",
|
| 889 |
+
"model.layers.35.mlp.up_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 890 |
+
"model.layers.35.mlp.up_proj.scales": "model-00003-of-00003.safetensors",
|
| 891 |
+
"model.layers.35.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 892 |
+
"model.layers.35.self_attn.k_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 893 |
+
"model.layers.35.self_attn.k_proj.qweight": "model-00003-of-00003.safetensors",
|
| 894 |
+
"model.layers.35.self_attn.k_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 895 |
+
"model.layers.35.self_attn.k_proj.scales": "model-00003-of-00003.safetensors",
|
| 896 |
+
"model.layers.35.self_attn.o_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 897 |
+
"model.layers.35.self_attn.o_proj.qweight": "model-00003-of-00003.safetensors",
|
| 898 |
+
"model.layers.35.self_attn.o_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 899 |
+
"model.layers.35.self_attn.o_proj.scales": "model-00003-of-00003.safetensors",
|
| 900 |
+
"model.layers.35.self_attn.q_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 901 |
+
"model.layers.35.self_attn.q_proj.qweight": "model-00003-of-00003.safetensors",
|
| 902 |
+
"model.layers.35.self_attn.q_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 903 |
+
"model.layers.35.self_attn.q_proj.scales": "model-00003-of-00003.safetensors",
|
| 904 |
+
"model.layers.35.self_attn.v_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 905 |
+
"model.layers.35.self_attn.v_proj.qweight": "model-00003-of-00003.safetensors",
|
| 906 |
+
"model.layers.35.self_attn.v_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 907 |
+
"model.layers.35.self_attn.v_proj.scales": "model-00003-of-00003.safetensors",
|
| 908 |
+
"model.layers.36.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 909 |
+
"model.layers.36.mlp.down_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 910 |
+
"model.layers.36.mlp.down_proj.qweight": "model-00003-of-00003.safetensors",
|
| 911 |
+
"model.layers.36.mlp.down_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 912 |
+
"model.layers.36.mlp.down_proj.scales": "model-00003-of-00003.safetensors",
|
| 913 |
+
"model.layers.36.mlp.gate_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 914 |
+
"model.layers.36.mlp.gate_proj.qweight": "model-00003-of-00003.safetensors",
|
| 915 |
+
"model.layers.36.mlp.gate_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 916 |
+
"model.layers.36.mlp.gate_proj.scales": "model-00003-of-00003.safetensors",
|
| 917 |
+
"model.layers.36.mlp.up_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 918 |
+
"model.layers.36.mlp.up_proj.qweight": "model-00003-of-00003.safetensors",
|
| 919 |
+
"model.layers.36.mlp.up_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 920 |
+
"model.layers.36.mlp.up_proj.scales": "model-00003-of-00003.safetensors",
|
| 921 |
+
"model.layers.36.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 922 |
+
"model.layers.36.self_attn.k_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 923 |
+
"model.layers.36.self_attn.k_proj.qweight": "model-00003-of-00003.safetensors",
|
| 924 |
+
"model.layers.36.self_attn.k_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 925 |
+
"model.layers.36.self_attn.k_proj.scales": "model-00003-of-00003.safetensors",
|
| 926 |
+
"model.layers.36.self_attn.o_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 927 |
+
"model.layers.36.self_attn.o_proj.qweight": "model-00003-of-00003.safetensors",
|
| 928 |
+
"model.layers.36.self_attn.o_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 929 |
+
"model.layers.36.self_attn.o_proj.scales": "model-00003-of-00003.safetensors",
|
| 930 |
+
"model.layers.36.self_attn.q_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 931 |
+
"model.layers.36.self_attn.q_proj.qweight": "model-00003-of-00003.safetensors",
|
| 932 |
+
"model.layers.36.self_attn.q_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 933 |
+
"model.layers.36.self_attn.q_proj.scales": "model-00003-of-00003.safetensors",
|
| 934 |
+
"model.layers.36.self_attn.v_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 935 |
+
"model.layers.36.self_attn.v_proj.qweight": "model-00003-of-00003.safetensors",
|
| 936 |
+
"model.layers.36.self_attn.v_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 937 |
+
"model.layers.36.self_attn.v_proj.scales": "model-00003-of-00003.safetensors",
|
| 938 |
+
"model.layers.37.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 939 |
+
"model.layers.37.mlp.down_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 940 |
+
"model.layers.37.mlp.down_proj.qweight": "model-00003-of-00003.safetensors",
|
| 941 |
+
"model.layers.37.mlp.down_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 942 |
+
"model.layers.37.mlp.down_proj.scales": "model-00003-of-00003.safetensors",
|
| 943 |
+
"model.layers.37.mlp.gate_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 944 |
+
"model.layers.37.mlp.gate_proj.qweight": "model-00003-of-00003.safetensors",
|
| 945 |
+
"model.layers.37.mlp.gate_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 946 |
+
"model.layers.37.mlp.gate_proj.scales": "model-00003-of-00003.safetensors",
|
| 947 |
+
"model.layers.37.mlp.up_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 948 |
+
"model.layers.37.mlp.up_proj.qweight": "model-00003-of-00003.safetensors",
|
| 949 |
+
"model.layers.37.mlp.up_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 950 |
+
"model.layers.37.mlp.up_proj.scales": "model-00003-of-00003.safetensors",
|
| 951 |
+
"model.layers.37.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 952 |
+
"model.layers.37.self_attn.k_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 953 |
+
"model.layers.37.self_attn.k_proj.qweight": "model-00003-of-00003.safetensors",
|
| 954 |
+
"model.layers.37.self_attn.k_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 955 |
+
"model.layers.37.self_attn.k_proj.scales": "model-00003-of-00003.safetensors",
|
| 956 |
+
"model.layers.37.self_attn.o_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 957 |
+
"model.layers.37.self_attn.o_proj.qweight": "model-00003-of-00003.safetensors",
|
| 958 |
+
"model.layers.37.self_attn.o_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 959 |
+
"model.layers.37.self_attn.o_proj.scales": "model-00003-of-00003.safetensors",
|
| 960 |
+
"model.layers.37.self_attn.q_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 961 |
+
"model.layers.37.self_attn.q_proj.qweight": "model-00003-of-00003.safetensors",
|
| 962 |
+
"model.layers.37.self_attn.q_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 963 |
+
"model.layers.37.self_attn.q_proj.scales": "model-00003-of-00003.safetensors",
|
| 964 |
+
"model.layers.37.self_attn.v_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 965 |
+
"model.layers.37.self_attn.v_proj.qweight": "model-00003-of-00003.safetensors",
|
| 966 |
+
"model.layers.37.self_attn.v_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 967 |
+
"model.layers.37.self_attn.v_proj.scales": "model-00003-of-00003.safetensors",
|
| 968 |
+
"model.layers.38.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 969 |
+
"model.layers.38.mlp.down_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 970 |
+
"model.layers.38.mlp.down_proj.qweight": "model-00003-of-00003.safetensors",
|
| 971 |
+
"model.layers.38.mlp.down_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 972 |
+
"model.layers.38.mlp.down_proj.scales": "model-00003-of-00003.safetensors",
|
| 973 |
+
"model.layers.38.mlp.gate_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 974 |
+
"model.layers.38.mlp.gate_proj.qweight": "model-00003-of-00003.safetensors",
|
| 975 |
+
"model.layers.38.mlp.gate_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 976 |
+
"model.layers.38.mlp.gate_proj.scales": "model-00003-of-00003.safetensors",
|
| 977 |
+
"model.layers.38.mlp.up_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 978 |
+
"model.layers.38.mlp.up_proj.qweight": "model-00003-of-00003.safetensors",
|
| 979 |
+
"model.layers.38.mlp.up_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 980 |
+
"model.layers.38.mlp.up_proj.scales": "model-00003-of-00003.safetensors",
|
| 981 |
+
"model.layers.38.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 982 |
+
"model.layers.38.self_attn.k_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 983 |
+
"model.layers.38.self_attn.k_proj.qweight": "model-00003-of-00003.safetensors",
|
| 984 |
+
"model.layers.38.self_attn.k_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 985 |
+
"model.layers.38.self_attn.k_proj.scales": "model-00003-of-00003.safetensors",
|
| 986 |
+
"model.layers.38.self_attn.o_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 987 |
+
"model.layers.38.self_attn.o_proj.qweight": "model-00003-of-00003.safetensors",
|
| 988 |
+
"model.layers.38.self_attn.o_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 989 |
+
"model.layers.38.self_attn.o_proj.scales": "model-00003-of-00003.safetensors",
|
| 990 |
+
"model.layers.38.self_attn.q_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 991 |
+
"model.layers.38.self_attn.q_proj.qweight": "model-00003-of-00003.safetensors",
|
| 992 |
+
"model.layers.38.self_attn.q_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 993 |
+
"model.layers.38.self_attn.q_proj.scales": "model-00003-of-00003.safetensors",
|
| 994 |
+
"model.layers.38.self_attn.v_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 995 |
+
"model.layers.38.self_attn.v_proj.qweight": "model-00003-of-00003.safetensors",
|
| 996 |
+
"model.layers.38.self_attn.v_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 997 |
+
"model.layers.38.self_attn.v_proj.scales": "model-00003-of-00003.safetensors",
|
| 998 |
+
"model.layers.39.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 999 |
+
"model.layers.39.mlp.down_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 1000 |
+
"model.layers.39.mlp.down_proj.qweight": "model-00003-of-00003.safetensors",
|
| 1001 |
+
"model.layers.39.mlp.down_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 1002 |
+
"model.layers.39.mlp.down_proj.scales": "model-00003-of-00003.safetensors",
|
| 1003 |
+
"model.layers.39.mlp.gate_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 1004 |
+
"model.layers.39.mlp.gate_proj.qweight": "model-00003-of-00003.safetensors",
|
| 1005 |
+
"model.layers.39.mlp.gate_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 1006 |
+
"model.layers.39.mlp.gate_proj.scales": "model-00003-of-00003.safetensors",
|
| 1007 |
+
"model.layers.39.mlp.up_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 1008 |
+
"model.layers.39.mlp.up_proj.qweight": "model-00003-of-00003.safetensors",
|
| 1009 |
+
"model.layers.39.mlp.up_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 1010 |
+
"model.layers.39.mlp.up_proj.scales": "model-00003-of-00003.safetensors",
|
| 1011 |
+
"model.layers.39.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
| 1012 |
+
"model.layers.39.self_attn.k_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 1013 |
+
"model.layers.39.self_attn.k_proj.qweight": "model-00003-of-00003.safetensors",
|
| 1014 |
+
"model.layers.39.self_attn.k_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 1015 |
+
"model.layers.39.self_attn.k_proj.scales": "model-00003-of-00003.safetensors",
|
| 1016 |
+
"model.layers.39.self_attn.o_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 1017 |
+
"model.layers.39.self_attn.o_proj.qweight": "model-00003-of-00003.safetensors",
|
| 1018 |
+
"model.layers.39.self_attn.o_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 1019 |
+
"model.layers.39.self_attn.o_proj.scales": "model-00003-of-00003.safetensors",
|
| 1020 |
+
"model.layers.39.self_attn.q_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 1021 |
+
"model.layers.39.self_attn.q_proj.qweight": "model-00003-of-00003.safetensors",
|
| 1022 |
+
"model.layers.39.self_attn.q_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 1023 |
+
"model.layers.39.self_attn.q_proj.scales": "model-00003-of-00003.safetensors",
|
| 1024 |
+
"model.layers.39.self_attn.v_proj.g_idx": "model-00003-of-00003.safetensors",
|
| 1025 |
+
"model.layers.39.self_attn.v_proj.qweight": "model-00003-of-00003.safetensors",
|
| 1026 |
+
"model.layers.39.self_attn.v_proj.qzeros": "model-00003-of-00003.safetensors",
|
| 1027 |
+
"model.layers.39.self_attn.v_proj.scales": "model-00003-of-00003.safetensors",
|
| 1028 |
+
"model.layers.4.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 1029 |
+
"model.layers.4.mlp.down_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1030 |
+
"model.layers.4.mlp.down_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1031 |
+
"model.layers.4.mlp.down_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1032 |
+
"model.layers.4.mlp.down_proj.scales": "model-00001-of-00003.safetensors",
|
| 1033 |
+
"model.layers.4.mlp.gate_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1034 |
+
"model.layers.4.mlp.gate_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1035 |
+
"model.layers.4.mlp.gate_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1036 |
+
"model.layers.4.mlp.gate_proj.scales": "model-00001-of-00003.safetensors",
|
| 1037 |
+
"model.layers.4.mlp.up_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1038 |
+
"model.layers.4.mlp.up_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1039 |
+
"model.layers.4.mlp.up_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1040 |
+
"model.layers.4.mlp.up_proj.scales": "model-00001-of-00003.safetensors",
|
| 1041 |
+
"model.layers.4.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 1042 |
+
"model.layers.4.self_attn.k_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1043 |
+
"model.layers.4.self_attn.k_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1044 |
+
"model.layers.4.self_attn.k_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1045 |
+
"model.layers.4.self_attn.k_proj.scales": "model-00001-of-00003.safetensors",
|
| 1046 |
+
"model.layers.4.self_attn.o_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1047 |
+
"model.layers.4.self_attn.o_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1048 |
+
"model.layers.4.self_attn.o_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1049 |
+
"model.layers.4.self_attn.o_proj.scales": "model-00001-of-00003.safetensors",
|
| 1050 |
+
"model.layers.4.self_attn.q_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1051 |
+
"model.layers.4.self_attn.q_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1052 |
+
"model.layers.4.self_attn.q_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1053 |
+
"model.layers.4.self_attn.q_proj.scales": "model-00001-of-00003.safetensors",
|
| 1054 |
+
"model.layers.4.self_attn.v_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1055 |
+
"model.layers.4.self_attn.v_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1056 |
+
"model.layers.4.self_attn.v_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1057 |
+
"model.layers.4.self_attn.v_proj.scales": "model-00001-of-00003.safetensors",
|
| 1058 |
+
"model.layers.5.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 1059 |
+
"model.layers.5.mlp.down_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1060 |
+
"model.layers.5.mlp.down_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1061 |
+
"model.layers.5.mlp.down_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1062 |
+
"model.layers.5.mlp.down_proj.scales": "model-00001-of-00003.safetensors",
|
| 1063 |
+
"model.layers.5.mlp.gate_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1064 |
+
"model.layers.5.mlp.gate_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1065 |
+
"model.layers.5.mlp.gate_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1066 |
+
"model.layers.5.mlp.gate_proj.scales": "model-00001-of-00003.safetensors",
|
| 1067 |
+
"model.layers.5.mlp.up_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1068 |
+
"model.layers.5.mlp.up_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1069 |
+
"model.layers.5.mlp.up_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1070 |
+
"model.layers.5.mlp.up_proj.scales": "model-00001-of-00003.safetensors",
|
| 1071 |
+
"model.layers.5.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 1072 |
+
"model.layers.5.self_attn.k_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1073 |
+
"model.layers.5.self_attn.k_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1074 |
+
"model.layers.5.self_attn.k_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1075 |
+
"model.layers.5.self_attn.k_proj.scales": "model-00001-of-00003.safetensors",
|
| 1076 |
+
"model.layers.5.self_attn.o_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1077 |
+
"model.layers.5.self_attn.o_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1078 |
+
"model.layers.5.self_attn.o_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1079 |
+
"model.layers.5.self_attn.o_proj.scales": "model-00001-of-00003.safetensors",
|
| 1080 |
+
"model.layers.5.self_attn.q_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1081 |
+
"model.layers.5.self_attn.q_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1082 |
+
"model.layers.5.self_attn.q_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1083 |
+
"model.layers.5.self_attn.q_proj.scales": "model-00001-of-00003.safetensors",
|
| 1084 |
+
"model.layers.5.self_attn.v_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1085 |
+
"model.layers.5.self_attn.v_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1086 |
+
"model.layers.5.self_attn.v_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1087 |
+
"model.layers.5.self_attn.v_proj.scales": "model-00001-of-00003.safetensors",
|
| 1088 |
+
"model.layers.6.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 1089 |
+
"model.layers.6.mlp.down_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1090 |
+
"model.layers.6.mlp.down_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1091 |
+
"model.layers.6.mlp.down_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1092 |
+
"model.layers.6.mlp.down_proj.scales": "model-00001-of-00003.safetensors",
|
| 1093 |
+
"model.layers.6.mlp.gate_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1094 |
+
"model.layers.6.mlp.gate_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1095 |
+
"model.layers.6.mlp.gate_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1096 |
+
"model.layers.6.mlp.gate_proj.scales": "model-00001-of-00003.safetensors",
|
| 1097 |
+
"model.layers.6.mlp.up_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1098 |
+
"model.layers.6.mlp.up_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1099 |
+
"model.layers.6.mlp.up_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1100 |
+
"model.layers.6.mlp.up_proj.scales": "model-00001-of-00003.safetensors",
|
| 1101 |
+
"model.layers.6.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 1102 |
+
"model.layers.6.self_attn.k_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1103 |
+
"model.layers.6.self_attn.k_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1104 |
+
"model.layers.6.self_attn.k_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1105 |
+
"model.layers.6.self_attn.k_proj.scales": "model-00001-of-00003.safetensors",
|
| 1106 |
+
"model.layers.6.self_attn.o_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1107 |
+
"model.layers.6.self_attn.o_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1108 |
+
"model.layers.6.self_attn.o_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1109 |
+
"model.layers.6.self_attn.o_proj.scales": "model-00001-of-00003.safetensors",
|
| 1110 |
+
"model.layers.6.self_attn.q_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1111 |
+
"model.layers.6.self_attn.q_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1112 |
+
"model.layers.6.self_attn.q_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1113 |
+
"model.layers.6.self_attn.q_proj.scales": "model-00001-of-00003.safetensors",
|
| 1114 |
+
"model.layers.6.self_attn.v_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1115 |
+
"model.layers.6.self_attn.v_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1116 |
+
"model.layers.6.self_attn.v_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1117 |
+
"model.layers.6.self_attn.v_proj.scales": "model-00001-of-00003.safetensors",
|
| 1118 |
+
"model.layers.7.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 1119 |
+
"model.layers.7.mlp.down_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1120 |
+
"model.layers.7.mlp.down_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1121 |
+
"model.layers.7.mlp.down_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1122 |
+
"model.layers.7.mlp.down_proj.scales": "model-00001-of-00003.safetensors",
|
| 1123 |
+
"model.layers.7.mlp.gate_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1124 |
+
"model.layers.7.mlp.gate_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1125 |
+
"model.layers.7.mlp.gate_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1126 |
+
"model.layers.7.mlp.gate_proj.scales": "model-00001-of-00003.safetensors",
|
| 1127 |
+
"model.layers.7.mlp.up_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1128 |
+
"model.layers.7.mlp.up_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1129 |
+
"model.layers.7.mlp.up_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1130 |
+
"model.layers.7.mlp.up_proj.scales": "model-00001-of-00003.safetensors",
|
| 1131 |
+
"model.layers.7.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 1132 |
+
"model.layers.7.self_attn.k_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1133 |
+
"model.layers.7.self_attn.k_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1134 |
+
"model.layers.7.self_attn.k_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1135 |
+
"model.layers.7.self_attn.k_proj.scales": "model-00001-of-00003.safetensors",
|
| 1136 |
+
"model.layers.7.self_attn.o_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1137 |
+
"model.layers.7.self_attn.o_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1138 |
+
"model.layers.7.self_attn.o_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1139 |
+
"model.layers.7.self_attn.o_proj.scales": "model-00001-of-00003.safetensors",
|
| 1140 |
+
"model.layers.7.self_attn.q_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1141 |
+
"model.layers.7.self_attn.q_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1142 |
+
"model.layers.7.self_attn.q_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1143 |
+
"model.layers.7.self_attn.q_proj.scales": "model-00001-of-00003.safetensors",
|
| 1144 |
+
"model.layers.7.self_attn.v_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1145 |
+
"model.layers.7.self_attn.v_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1146 |
+
"model.layers.7.self_attn.v_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1147 |
+
"model.layers.7.self_attn.v_proj.scales": "model-00001-of-00003.safetensors",
|
| 1148 |
+
"model.layers.8.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 1149 |
+
"model.layers.8.mlp.down_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1150 |
+
"model.layers.8.mlp.down_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1151 |
+
"model.layers.8.mlp.down_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1152 |
+
"model.layers.8.mlp.down_proj.scales": "model-00001-of-00003.safetensors",
|
| 1153 |
+
"model.layers.8.mlp.gate_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1154 |
+
"model.layers.8.mlp.gate_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1155 |
+
"model.layers.8.mlp.gate_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1156 |
+
"model.layers.8.mlp.gate_proj.scales": "model-00001-of-00003.safetensors",
|
| 1157 |
+
"model.layers.8.mlp.up_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1158 |
+
"model.layers.8.mlp.up_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1159 |
+
"model.layers.8.mlp.up_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1160 |
+
"model.layers.8.mlp.up_proj.scales": "model-00001-of-00003.safetensors",
|
| 1161 |
+
"model.layers.8.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 1162 |
+
"model.layers.8.self_attn.k_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1163 |
+
"model.layers.8.self_attn.k_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1164 |
+
"model.layers.8.self_attn.k_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1165 |
+
"model.layers.8.self_attn.k_proj.scales": "model-00001-of-00003.safetensors",
|
| 1166 |
+
"model.layers.8.self_attn.o_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1167 |
+
"model.layers.8.self_attn.o_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1168 |
+
"model.layers.8.self_attn.o_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1169 |
+
"model.layers.8.self_attn.o_proj.scales": "model-00001-of-00003.safetensors",
|
| 1170 |
+
"model.layers.8.self_attn.q_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1171 |
+
"model.layers.8.self_attn.q_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1172 |
+
"model.layers.8.self_attn.q_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1173 |
+
"model.layers.8.self_attn.q_proj.scales": "model-00001-of-00003.safetensors",
|
| 1174 |
+
"model.layers.8.self_attn.v_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1175 |
+
"model.layers.8.self_attn.v_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1176 |
+
"model.layers.8.self_attn.v_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1177 |
+
"model.layers.8.self_attn.v_proj.scales": "model-00001-of-00003.safetensors",
|
| 1178 |
+
"model.layers.9.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 1179 |
+
"model.layers.9.mlp.down_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1180 |
+
"model.layers.9.mlp.down_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1181 |
+
"model.layers.9.mlp.down_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1182 |
+
"model.layers.9.mlp.down_proj.scales": "model-00001-of-00003.safetensors",
|
| 1183 |
+
"model.layers.9.mlp.gate_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1184 |
+
"model.layers.9.mlp.gate_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1185 |
+
"model.layers.9.mlp.gate_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1186 |
+
"model.layers.9.mlp.gate_proj.scales": "model-00001-of-00003.safetensors",
|
| 1187 |
+
"model.layers.9.mlp.up_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1188 |
+
"model.layers.9.mlp.up_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1189 |
+
"model.layers.9.mlp.up_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1190 |
+
"model.layers.9.mlp.up_proj.scales": "model-00001-of-00003.safetensors",
|
| 1191 |
+
"model.layers.9.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
| 1192 |
+
"model.layers.9.self_attn.k_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1193 |
+
"model.layers.9.self_attn.k_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1194 |
+
"model.layers.9.self_attn.k_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1195 |
+
"model.layers.9.self_attn.k_proj.scales": "model-00001-of-00003.safetensors",
|
| 1196 |
+
"model.layers.9.self_attn.o_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1197 |
+
"model.layers.9.self_attn.o_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1198 |
+
"model.layers.9.self_attn.o_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1199 |
+
"model.layers.9.self_attn.o_proj.scales": "model-00001-of-00003.safetensors",
|
| 1200 |
+
"model.layers.9.self_attn.q_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1201 |
+
"model.layers.9.self_attn.q_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1202 |
+
"model.layers.9.self_attn.q_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1203 |
+
"model.layers.9.self_attn.q_proj.scales": "model-00001-of-00003.safetensors",
|
| 1204 |
+
"model.layers.9.self_attn.v_proj.g_idx": "model-00001-of-00003.safetensors",
|
| 1205 |
+
"model.layers.9.self_attn.v_proj.qweight": "model-00001-of-00003.safetensors",
|
| 1206 |
+
"model.layers.9.self_attn.v_proj.qzeros": "model-00001-of-00003.safetensors",
|
| 1207 |
+
"model.layers.9.self_attn.v_proj.scales": "model-00001-of-00003.safetensors",
|
| 1208 |
+
"model.norm.weight": "model-00003-of-00003.safetensors"
|
| 1209 |
+
}
|
| 1210 |
+
}
|
special_tokens_map.json
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bos_token": {
|
| 3 |
+
"content": "<s>",
|
| 4 |
+
"lstrip": false,
|
| 5 |
+
"normalized": false,
|
| 6 |
+
"rstrip": false,
|
| 7 |
+
"single_word": false
|
| 8 |
+
},
|
| 9 |
+
"eos_token": {
|
| 10 |
+
"content": "</s>",
|
| 11 |
+
"lstrip": false,
|
| 12 |
+
"normalized": false,
|
| 13 |
+
"rstrip": false,
|
| 14 |
+
"single_word": false
|
| 15 |
+
},
|
| 16 |
+
"pad_token": {
|
| 17 |
+
"content": "<pad>",
|
| 18 |
+
"lstrip": false,
|
| 19 |
+
"normalized": false,
|
| 20 |
+
"rstrip": false,
|
| 21 |
+
"single_word": false
|
| 22 |
+
},
|
| 23 |
+
"unk_token": {
|
| 24 |
+
"content": "<unk>",
|
| 25 |
+
"lstrip": false,
|
| 26 |
+
"normalized": false,
|
| 27 |
+
"rstrip": false,
|
| 28 |
+
"single_word": false
|
| 29 |
+
}
|
| 30 |
+
}
|
tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b0240ce510f08e6c2041724e9043e33be9d251d1e4a4d94eb68cd47b954b61d2
|
| 3 |
+
size 17078292
|
tokenizer_config.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|