File size: 87,184 Bytes
d493a02
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
---
tags:
- sentence-transformers
- sentence-similarity
- feature-extraction
- dense
- generated_from_trainer
- dataset_size:810
- loss:MultipleNegativesRankingLoss
base_model: sentence-transformers/paraphrase-multilingual-mpnet-base-v2
widget:
- source_sentence: CBRN defense, predictive analytics, natural language understanding
  sentences:
  - "experience with speech interfaces Lead and evaluate changing dialog evaluation\
    \ conventions, test tooling developments, and pilot processes to support expansion\
    \ to new data areas Continuously evaluate workflow tools and processes and offer\
    \ solutions to ensure they are efficient, high quality, and scalable Provide expert\
    \ support for a large and growing team of data analysts Provide support for ongoing\
    \ and new data collection efforts as a subject matter expert on conventions and\
    \ use of the data Conduct research studies to understand speech and customer-Alexa\
    \ interactions Assist scientists, program and product managers, and other stakeholders\
    \ in defining and validating customer experience metrics\n\nWe are open to hiring\
    \ candidates to work out of one of the following locations:\n\nBoston, MA, USA\
    \ | Seattle, WA, USA\n\nBasic Qualifications\n\n 3+ years of data querying languages\
    \ (e.g. SQL), scripting languages (e.g. Python) or statistical/mathematical software\
    \ (e.g. R, SAS, Matlab, etc.) experience 2+ years of data scientist experience\
    \ Bachelor's degree Experience applying theoretical models in an applied environment\n\
    \nPreferred Qualifications\n\n Experience in Python, Perl, or another scripting\
    \ language Experience in a ML or data scientist role with a large technology company\
    \ Master's degree in a quantitative field such as statistics, mathematics, data\
    \ science, business analytics, economics, finance, engineering, or computer science\n\
    \nAmazon is committed to a diverse and inclusive workplace. Amazon is \n\nOur\
    \ compensation reflects the cost of labor across several US geographic markets.\
    \ The base pay for this position ranges from $111,600/year in our lowest geographic\
    \ market up to $212,800/year in our highest geographic market. Pay is based on\
    \ a number of factors including market location and may vary depending on job-related\
    \ knowledge, skills, and experience. Amazon is a total compensation company. Dependent\
    \ on the position offered, equity, sign-on payments, and other forms of compensation\
    \ may be provided as part of a total compensation package, in addition to a full\
    \ range of medical, financial, and/or other benefits. For more information, please\
    \ visit https://www.aboutamazon.com/workplace/employee-benefits. This position\
    \ will remain posted until filled. Applicants should apply via our internal or\
    \ external career site.\n\n\nCompany - Amazon.com Services LLC\n\nJob ID: A2610750"
  - 'Skills: Your Expertise:

    5+ years in industry experience and a degree (Masters or PhD is a plus) in a quantitative
    field (e.g., Statistics, Econometrics, Computer Science, Engineering, Mathematics,
    Data Science, Operations Research).Expert communication and collaboration skills
    with the ability to work effectively with internal teams in a cross-cultural and
    cross-functional environment. Ability to conduct rigorous analysis and communicate
    conclusions to both technical and non-technical audiencesExperience partnering
    with internal teams to drive action and providing expertise and direction on analytics,
    data science, experimental design, and measurementExperience in analysis of A|B
    experiments and statistical data analysisExperience designing and building metrics,
    from conception to building prototypes with data pipelinesStrong knowledge in
    at least one programming language (Python or R) and in SQLAbility to drive data
    strategies, with a central source of truth to impact business decisionsKnowledge
    and experience in insurance industry - a plusKnowledge and experience in customer
    experience measurement - a plus

    Keywords:Education: Minimum: BS/BA in CS or related field (or self-taught/ equivalent
    work experience) Preferred: MS/MA in CS or related field'
  - "requirements of the program or company.\n\n Working across the globe, V2X builds\
    \ smart solutions designed to integrate physical and digital infrastructure from\
    \ base to battlefield. We bring 120 years of successful mission support to improve\
    \ security, streamline logistics, and enhance readiness. Aligned around a shared\
    \ purpose, our $3.9B company and 16,000 people work alongside our clients, here\
    \ and abroad, to tackle their most complex challenges with integrity, respect,\
    \ responsibility, and professionalism. \n\nAt V2X, we are making a difference\
    \ by delivering decision support tools critical for the protection of our forces\
    \ when threatened by both physical and Chemical, Biological, Radiological, or\
    \ Nuclear (CBRN) threats.\n\nWe are expanding in data science to provide the best\
    \ information possible utilizing the latest techniques in Machine Learning (including\
    \ Deep Learning, Neural network). We are on the forefront of CBRN defense and\
    \ we are looking for talented Data Scientists that have applied experience in\
    \ the fields of artificial intelligence, machine learning and/or natural language\
    \ processing to join our team. Our data scientists work closely everyday with\
    \ project managers, subject matter experts and software engineers to work on challenges\
    \ in machine intelligence, data mining, and machine learning, and work together\
    \ with agility to build capabilities that impress our customers. We partner and\
    \ collaborate with universities to being best minds together.\n\nData scientists\
    \ will have opportunities to work on projects with highest priority to our business.\
    \ Vital to success in this role is the ability to determine, define and deploy\
    \ predictive / prescriptive analytic solutions to identify and perform root cause\
    \ analysis on adverse trends, by choosing best fit methods, defining algorithms,\
    \ and validating and deploying models to achieve results.\n\nResponsibilities\n\
    \nMajor Job Activities:\n\n Partner with our development teams to solve problems\
    \ and identify trends and opportunities to leverage data from multiple sources.\
    \  Collaborate across multiple teams.  Passionate about working with large and\
    \ complex unstructured and structured data sets.  Strong communication and interpersonal\
    \ skills. You should be able to work across functions and effectively present,\
    \ recommend and communicate a position by demonstrating its value and tradeoffs.\
    \  Comfortable conducting design, algorithm, and code reviews.  Able to self-direct\
    \ and succeed with minimal guidance. \n\nMaterial & Equipment Directly Used:\n\
    \nComputer, Phone, and basic office materials.\n\nWorking Environment:\n\n Function\
    \ in an office environment in a stationary position approximately 50 percent of\
    \ the time or more.  Must be able to operate standard office equipment, such as\
    \ a computer, copy machine, and printer. \n\nQualifications\n\nEducation / Certifications:\n\
    \n Bachelor’s degree in a computer, engineering, or quantitative discipline (e.g.,\
    \ statistics, operations research, bioinformatics, economics, computational biology,\
    \ computer science, mathematics, physics, electrical engineering, industrial engineering).\
    \  Master's or Ph.D. in a quantitative discipline preferred. \n\nClearance Requirement:\
    \ \n\nMust have or be able to obtain an active U.S. DoD Secret (or higher) level\
    \ Security Clearance.\n\nExperience / Skills:\n\n 5+ years of relevant work experience\
    \ in data analysis or related field. (e.g., statistician, data analyst, data scientist).\
    \  Programming experience in one or more of the following: R, MATLAB, C, C++,\
    \ Java, Python, Scala  Experience in Natural Language Understanding, Computer\
    \ Vision, Machine Learning, Algorithmic Foundations of Optimization, Data Mining\
    \ or Machine Intelligence (Artificial Intelligence).  Experience with statistical\
    \ software (e.g., R, Octave, Julia, MATLAB, pandas) and database languages (e.g.,\
    \ SQL).  Experience with machine learning related open source libraries including,\
    \ but not limited to: Hadoop, Spark, SciKit-Learn, TensorFlow, etc.  Contribution\
    \ to research communities and/or efforts, including publishing papers at conferences.\
    \ \n\nWe are committed to an inclusive and diverse workplace that values and supports\
    \ the contributions of each individual. This commitment along with our common\
    \ Vision and Values of Integrity, Respect, and Responsibility, allows us to leverage\
    \ differences, encourage innovation and expand our success in the global marketplace.\
    \ V2X is an Equal Opportunity /Affirmative Action Employer. All qualified applicants\
    \ will receive consideration for employment without regard to race, color, religion,\
    \ age, sex, national origin, protected veteran status or status as an individual\
    \ with a disability."
- source_sentence: Senior Data Analyst Pricing, data product automation, pricing strategy
    analysis
  sentences:
  - Skills You BringBachelor’s or Master’s Degree in a technology related field (e.g.
    Engineering, Computer Science, etc.) required with 6+ years of experienceInformatica
    Power CenterGood experience with ETL technologiesSnaplogicStrong SQLProven data
    analysis skillsStrong data modeling skills doing either Dimensional or Data Vault
    modelsBasic AWS Experience Proven ability to deal with ambiguity and work in fast
    paced environmentExcellent interpersonal and communication skillsExcellent collaboration
    skills to work with multiple teams in the organization
  - "experience, an annualized transactional volume of $140 billion in 2023, and approximately\
    \ 3,200 employees located in 12+ countries, Paysafe connects businesses and consumers\
    \ across 260 payment types in over 40 currencies around the world. Delivered through\
    \ an integrated platform, Paysafe solutions are geared toward mobile-initiated\
    \ transactions, real-time analytics and the convergence between brick-and-mortar\
    \ and online payments. Further information is available at www.paysafe.com.\n\n\
    Are you ready to make an impact? Join our team that is inspired by a unified vision\
    \ and propelled by passion.\n\nPosition Summary\n\nWe are looking for a dynamic\
    \ and flexible, Senior Data Analyst, Pricing to support our global Sales and Product\
    \ organizations with strategic planning, analysis, and commercial pricing efforts\
    \ . As a Senior Data Analyst , you will be at the frontier of building our Pricing\
    \ function to drive growth through data and AI-enabled capabilities. This opportunity\
    \ is high visibility for someone hungry to drive the upward trajectory of our\
    \ business and be able to contribute to their efforts in the role in our success.\n\
    \nYou will partner with Product Managers to understand their commercial needs,\
    \ then prioritize and work with a cross-functional team to deliver pricing strategies\
    \ and analytics-based solutions to solve and execute them. Business outcomes will\
    \ include sustainable growth in both revenues and gross profit.\n\nThis role is\
    \ based in Jacksonville, Florida and offers a flexible hybrid work environment\
    \ with 3 days in the office and 2 days working remote during the work week.\n\n\
    Responsibilities\n\n Build data products that power the automation and effectiveness\
    \ of our pricing function, driving better quality revenues from merchants and\
    \ consumers.  Partner closely with pricing stakeholders (e.g., Product, Sales,\
    \ Marketing) to turn raw data into actionable insights. Help ask the right questions\
    \ and find the answers.  Dive into complex pricing and behavioral data sets, spot\
    \ trends and make interpretations.  Utilize modelling and data-mining skills to\
    \ find new insights and opportunities.  Turn findings into plans for new data\
    \ products or visions for new merchant features.  Partner across merchant Product,\
    \ Sales, Marketing, Development and Finance to build alignment, engagement and\
    \ excitement for new products, features and initiatives.  Ensure data quality\
    \ and integrity by following and enforcing data governance policies, including\
    \ alignment on data language. \n\n  Qualifications  \n\n Bachelor’s degree in\
    \ a related field of study (Computer Science, Statistics, Mathematics, Engineering,\
    \ etc.) required.  5+ years of experience of in-depth data analysis role, required;\
    \ preferably in pricing context with B2B & B2C in a digital environment.  Proven\
    \ ability to visualize data intuitively, cleanly and clearly in order to make\
    \ important insights simplified.  Experience across large and complex datasets,\
    \ including customer behavior, and transactional data.  Advanced in SQL and in\
    \ Python, preferred.  Experience structuring and analyzing A/B tests, elasticities\
    \ and interdependencies, preferred.  Excellent communication and presentation\
    \ skills, with the ability to explain complex data insights to non-technical audiences.\
    \ \n\n Life at Paysafe: \n\nOne network. One partnership. At Paysafe, this is\
    \ not only our business model; this is our mindset when it comes to our team.\
    \ Being a part of Paysafe means you’ll be one of over 3,200 members of a world-class\
    \ team that drives our business to new heights every day and where we are committed\
    \ to your personal and professional growth.\n\nOur culture values humility, high\
    \ trust & autonomy, a desire for excellence and meeting commitments, strong team\
    \ cohesion, a sense of urgency, a desire to learn, pragmatically pushing boundaries,\
    \ and accomplishing goals that have a direct business impact.\n\n \n\nPaysafe\
    \ provides equal employment opportunities to all employees, and applicants for\
    \ employment, and prohibits discrimination of any type concerning ethnicity, religion,\
    \ age, sex, national origin, disability status, sexual orientation, gender identity\
    \ or expression, or any other protected characteristics. This policy applies to\
    \ all terms and conditions of recruitment and employment. If you need any reasonable\
    \ adjustments, please let us know. We will be happy to help and look forward to\
    \ hearing from you."
  - "Experience : 10 yearsLocation : RemoteDuration: Full TimeJob DetailsData Warehouse,\
    \ ETL, Advanced SQL,Data Profiling, Source to Target Mapping,Business Requirement\
    \ Document, FRS, Healthcare.Should be able to navigate the code - developer background\n\
    \uFEFFThanks & Regard's\nMohd FurquanLead Technical RecruiterE-mail: furqan@msrcosmos.comDirect\
    \ No: +1 925 313 8949LinkedIn-ID :linkedin.com/in/mohd-furquan-94237816aVisit\
    \ us: www.msrcosmos.com"
- source_sentence: CPG data analysis, Nielsen IRI expertise, Power Query dashboard
    development
  sentences:
  - 'Skills :

    a) Azure Data Factory – Min 3 years of project experiencea. Design of pipelinesb.
    Use of project with On-prem to Cloud Data Migrationc. Understanding of ETLd. Change
    Data Capture from Multiple Sourcese. Job Schedulingb) Azure Data Lake – Min 3
    years of project experiencea. All steps from design to deliverb. Understanding
    of different Zones and design principalc) Data Modeling experience Min 5 Yearsa.
    Data Mart/Warehouseb. Columnar Data design and modelingd) Reporting using PowerBI
    Min 3 yearsa. Analytical Reportingb. Business Domain Modeling and data dictionary

    Interested please apply to the job, looking only for W2 candidates.'
  - "experienced and highly skilled Sr Data Engineer to join us. This role requires\
    \ a seasoned professional with a deep understanding of automated data pipelines,\
    \ cloud infrastructure, databases, and workflow engines. The ideal candidate will\
    \ have a minimum of 5 years of technical lead experience in the medical device\
    \ field and at least 7 years of experience in data engineering. Proficiency in\
    \ Python and a proven track record of leading projects to completion are essential.\n\
    \nPrimary Duties\n\nDesign, develop, and manage robust, secure, scalable, highly\
    \ available, and dynamic solutions to drive business objectives. Lead the architecture\
    \ and implementation of advanced cloud-based data engineering solutions, leveraging\
    \ AWS technologies and best practices. Manage and optimize data pipelines, ensuring\
    \ timely and accurate data availability for analytics and machine learning applications.\
    \ Oversee the administration and performance tuning of databases and workflow\
    \ engines. Collaborate with cross-functional teams (e.g., product management,\
    \ IT, software engineering) to define data requirements, integrate systems, and\
    \ implement data governance and security policies. Mentor junior data engineers\
    \ and oversee the team's development efforts, promoting best practices in coding,\
    \ architecture, and data management. Stay abreast of emerging technologies and\
    \ trends in data engineering, cloud services, and the medical device industry\
    \ to drive innovation and competitive advantage. \n\nKnowledge, Experience & Skills\n\
    \nDegree in Computer Science, Engineering, Information Systems, or a related field.\
    \ Requiring a minimum of Bachelor’s degree +7yrs of experience or a Master’s degree\
    \ +5yrs of experience. Minimum of 7 years of experience in data engineering, with\
    \ expertise in developing and managing automated data pipelines, AWS cloud infrastructure,\
    \ databases, and workflow engines. Certifications in AWS and data engineering\
    \ preferred. Experience with machine learning algorithms and data modeling techniques.\
    \ At least 5 years of experience in the medical device IVD industry, with a strong\
    \ understanding of FDA regulatory standards and compliance requirements. Expert\
    \ proficiency in Python programming and software engineering principles. Demonstrated\
    \ experience with AWS services (e.g., EC2, RDS, S3, Lambda, Glue, Redshift, Athena,\
    \ EMR) and data pipeline tools (e.g., Apache Airflow, Luigi, etc). Strong knowledge\
    \ of database management (Postgres and Snowflake), SQL, and NoSQL databases. Adept\
    \ at queries, report writing and presenting findings Experienced in developing\
    \ and maintaining ETL pipelines in a cloud environmentExperienced in Unit Testing\
    \ preferred Strong analytical skills with the ability to organize, analyze, and\
    \ disseminate information with attention to detail and accuracy Excellent communication\
    \ and task management skills. Comfort working in a dynamic, fast-paced, research-oriented\
    \ group with several ongoing concurrent projectsFull fluency (verbal and written)\
    \ of the English language is a must. \n\nThe estimated salary range for this role\
    \ based in California is between $148,700 and $178,400 annually. This role is\
    \ eligible to receive a variable annual bonus based on company, team, and individual\
    \ performance per bioMerieux’s bonus program. This range may differ from ranges\
    \ offered for similar positions elsewhere in the country given differences in\
    \ cost of living. Actual compensation within this range is determined based on\
    \ the successful candidate’s experience and will be presented in writing at the\
    \ time of the offer.\n\nIn addition, bioMérieux offers a competitive Total Rewards\
    \ package that may include:\n\nA choice of medical (including prescription), dental,\
    \ and vision plans providing nationwide coverage and telemedicine optionsCompany-Provided\
    \ Life and Accidental Death InsuranceShort and Long-Term Disability InsuranceRetirement\
    \ Plan including a generous non-discretionary employer contribution and employer\
    \ match. Adoption AssistanceWellness ProgramsEmployee Assistance ProgramCommuter\
    \ BenefitsVarious voluntary benefit offeringsDiscount programsParental leaves\n\
    \nBioFire Diagnostics, LLC. is an Equal Opportunity/Affirmative Action Employer.\
    \ All qualified applicants will receive consideration for employment without regard\
    \ to race, color, religion, sex, sexual orientation, gender identity, national\
    \ origin, age, protected veteran or disabled status, or genetic information.\n\
    \nPlease be advised that the receipt of satisfactory responses to reference requests\
    \ and the provision of satisfactory proof of an applicant’s identity and legal\
    \ authorization to work in the United States are required of all new hires. Any\
    \ misrepresentation, falsification, or material omission may result in the failure\
    \ to receive an offer, the retraction of an offer, or if already hired, dismissal.\
    \ If you are a qualified individual with a disability, you may request a reasonable\
    \ accommodation in BioFire Diagnostics’ application process by contacting us via\
    \ telephone at (385) 770-1132, by email at [email protected], or by dialing 711\
    \ for access to Telecommunications Relay Services (TRS)."
  - "requirements into analytical frameworks.Dashboard Development: Design and maintain\
    \ dashboards using Power Query in Excel, good in analytics in generating metrics\
    \ & measures and ensuring accurate and real-time data representation. \nRequired\
    \ QualificationsProfessional Experience: 3-6 years as a business analyst, with\
    \ mandatory experience in the CPG sector and should have worked on brand dataTechnical\
    \ Proficiency: Advanced skills in Excel and Power Query;Communication Skills:\
    \ Exceptional ability to communicate complex data insights to non-technical stakeholders.Location:\
    \ Position based in Springdale. Preferred AttributesProven experience in data-driven\
    \ decision-making processes.Ability to handle multiple projects simultaneously,\
    \ with a focus on deadlines and results."
- source_sentence: ETL Pipelines, Apache Spark, AirFlow
  sentences:
  - "Qualifications\n\n - Currently enrolled in a Bachelor’s or Master’s degree in\
    \ Software Development, Computer Science, Computer Engineering, or a related technical\
    \ discipline\n- Must obtain work authorization in country of employment at the\
    \ time of hire, and maintain ongoing work authorization during employment.\n\n\
    Preferred Qualifications: \n- Fluency in SQL or other programming languages (Python,\
    \ R etc) for data manipulation\n- Ability to thrive in a fast paced work environment\
    \ \n- Ability to drive projects to completion with minimal guidance\n- Ability\
    \ to communicate the results of analyses in a clear and effective manner\n\nTikTok\
    \ is committed to creating an inclusive space where employees are valued for their\
    \ skills, experiences, and unique perspectives. Our platform connects people from\
    \ across the globe and so does our workplace. At TikTok, our mission is to inspire\
    \ creativity and bring joy. To achieve that goal, we are committed to celebrating\
    \ our diverse voices and to creating an environment that reflects the many communities\
    \ we reach. We are passionate about this and hope you are too.\n\nTikTok is committed\
    \ to providing reasonable accommodations in our recruitment processes for candidates\
    \ with disabilities, pregnancy, sincerely held religious beliefs or other reasons\
    \ protected by applicable laws. If you need assistance or a reasonable accommodation,\
    \ please reach out to us at https://shorturl.at/cdpT2\n\nBy submitting an application\
    \ for this role, you accept and agree to our global applicant privacy policy,\
    \ which may be accessed here: https://careers.tiktok.com/legal/privacy. \n\nJob\
    \ Information:\n\n【For Pay Transparency】Compensation Description (annually) The\
    \ base salary range for this position in the selected city is $45 - $45annually.\
    \ We cover 100% premium coverage for Full-Time intern medical insurance after\
    \ 90 days from the date of hire. Medical coverage only, no dental or vision coverage.Our\
    \ time off and leave plans are: Paid holidays and paid sick leave. The sick leave\
    \ entitlement is based on the time you join.We also provide mental and emotional\
    \ health benefits through our Employee Assistance Program and provide reimbursements\
    \ for your mobile phone expense. The Company reserves the right to modify or change\
    \ these benefits programs at any time, with or without notice."
  - "Experience as a Product Data Analyst at TGG:Achieving business results as a client\
    \ facing consultant for our clients in various types of engagements within a variety\
    \ of industries.Delivering high quality work to our clients within our technology\
    \ service line. Being part of a collaborative, values-based firm that has a reputation\
    \ for great work and satisfied clients.Working with senior IT leaders to communicate\
    \ strategic goals to their organization, including leading client and internal\
    \ development teams on best practices.\nWhat You Will Work On:Analyze large datasets\
    \ to identify patterns, trends, and opportunities for product optimization.Develop\
    \ and maintain dashboards and reports to track key performance metrics.Collaborate\
    \ with product managers, marketers, and engineers to ideate, prioritize, and implement\
    \ data-driven initiatives.Conduct A/B testing and other statistical analyses to\
    \ evaluate the effectiveness of product changes.Communicate findings and recommendations\
    \ to stakeholders through clear and concise presentations.Contribute analytical\
    \ insights to inform product vision and deliver value.\nWho Will You Work With:Client\
    \ stakeholders ranging from individual contributors to senior executives.A collaborative\
    \ team of consultants that deliver outstanding client service.TGG partners, principals,\
    \ account leaders, managers, and staff supporting you to excel within client projects\
    \ and to achieve your professional development goals.\nExamples of What You Bring\
    \ to the Table:You have strong analysis capabilities and thrive on working collaboratively\
    \ to deliver successful results for clients. You have experience with these technologies:Proficiency\
    \ in SQL and Python for data extraction, manipulation, and analysis.Strong understanding\
    \ of statistical concepts and techniques.Intermediate experience with Tableau,\
    \ Power BI, Adobe Analytics, or similar BI tools.Ability to analyze requirements,\
    \ design, implement, debug, and deploy Cloud Platform services and components.At\
    \ least basic exposure to data science and machine learning methods.Familiarity\
    \ with source control best practices: Define, Setup/Configure, Deploy and Maintain\
    \ source code (e.g. GIT, VisualSafe Source).Ability to develop and schedule processes\
    \ to extract, transform, and store data from these systems: SQL databases, Azure\
    \ cloud services, Google cloud service, Snowflake.4-8 years of relevant experience.Bachelor’s\
    \ degree in Computer Science, Statistics, Economics, Mathematics, or a related\
    \ field; or equivalent combination of education, training, and experience.Analytical\
    \ Product Mindset: Ability to approach problems analytically and derive actionable\
    \ insights from complex datasets, while remaining focused on providing value to\
    \ customers Strategic Thinking: Demonstrated ability to translate data findings\
    \ into strategic, achievable recommendations to drive business outcomes.Communication\
    \ Skills: Excellent verbal and written communication skills.Ability to effectively\
    \ convey technical concepts from technical to non-technical stakeholders and vice-versa.Team\
    \ Player: Proven track record of collaborating effectively with cross-functional\
    \ teams in a fast-paced environment.Adaptability: Have consistently demonstrated\
    \ the ability to bring structure to complex, unstructured environments.Familiarity\
    \ with Agile development methodologies.Ability to adapt to changing priorities\
    \ to thrive in dynamic work environments.\nSalary and Benefits:Nothing is more\
    \ important to us than the well-being of our team. That is why we are proud to\
    \ offer a full suite of competitive health benefits along with additional benefits\
    \ such as: flexible PTO, a professional development stipend and work from home\
    \ stipend, volunteer opportunities, and team social activities.\nSalaries vary\
    \ and are dependent on considerations such as: experience and specific skills/certifications.\
    \ The base plus target bonus total compensation range for this role is $95,000\
    \ - $125,000. Additional compensation beyond this range is available as a result\
    \ of leadership and business development opportunities. Salary details are discussed\
    \ openly during the hiring process. \nWork Environment:TGG is headquartered in\
    \ Portland, Oregon, and has team members living in various locations across the\
    \ United States. Our consultants must have the ability to travel and to work remotely\
    \ or onsite. Each engagement has unique conditions, and we work collaboratively\
    \ to meet both our client and team's needs regarding onsite and travel requirements.\
    \ \nWhy The Gunter Group:TGG was created to be different, to be relational, to\
    \ be insightful, and to maximize potential for our consultants, our clients, and\
    \ our community. We listen first so we can learn, analyze, and deliver meaningful\
    \ solutions for our clients. Our compass points towards our people and our “Non-Negotiables”\
    \ always. Our driven employees make us who we are — a talented team of leaders\
    \ with deep and diverse professional experience.If you think this role is the\
    \ right fit, please submit your resume and cover letter so we can learn more about\
    \ you. \nThe Gunter Group LLC is"
  - 'Requirements & Day-to-Day:  Design, develop, and support scalable data processing
    pipelines using Apache Spark and Java/Scala. Lead a talented team and make a significant
    impact on our data engineering capabilities. Implement and manage workflow orchestration
    with AirFlow for efficient data processing. Proficiently use SQL for querying
    and data manipulation tasks. Collaborate with cross-functional teams to gather
    requirements and ensure alignment with data engineering solutions.  Essential
    Criteria:  a bachelor’s degree in computer science or another relevant discipline,
    and a minimum of five years of relevant experience in data engineering. Solid
    experience with Apache Spark for large-scale data processing. Proficiency in Java
    or Scala programming languages. Strong knowledge of AirFlow for workflow orchestration.
    Proficient in SQL for data querying and manipulation.'
- source_sentence: Data organization, document analysis, records management
  sentences:
  - "skills and build your career in a rapidly evolving business climate? Are you\
    \ looking for a career where professional development is embedded in your employer’s\
    \ core culture? If so, Chenega Military, Intelligence & Operations Support (MIOS)\
    \ could be the place for you! Join our team of professionals who support large-scale\
    \ government operations by leveraging cutting-edge technology and take your career\
    \ to the next level!\n\nAs one of the newest Chenega companies, Chenega Defense\
    \ & Aerospace Solutions (CDAS) was developed with the purpose of providing expert\
    \ Engineering and Technical Support Services to federal customers.\n\nThe Data\
    \ Analyst will analyze a large variety of documents to ensure proper placement\
    \ in physical files, perform high-level scanning of master file documents to convert\
    \ them into an electronic format, and provide meticulous organization and management\
    \ of case files, including sorting and categorizing documents before scanning.\n\
    \nResponsibilities\n\nWork within the Standard Operating Procedure for the organization\
    \ of physical files containing documents of various types Establish or maintain\
    \ physical files, including proper placement of documents as they are createdDisseminate\
    \ significant amounts of information with attention to detail and accuracyPerform\
    \ word processing tasksPerform data entry and metadata entry for electronic documentsReconcile\
    \ inconsistenciesGather information and organize investigative packages, case\
    \ files, or presentationsObtain additional information from other investigative\
    \ agencies or databasesVerify information and files against the tracking systemMaintain\
    \ internal status information on the disposition of designated information and\
    \ filesDistribute and receive documentsAssist analyst or government official in\
    \ obtaining or collecting all documents or information to complete case fileProvide\
    \ administrative information and assistance concerning the case or files to other\
    \ agencies or organizationsOther duties as assigned\n\n\nQualifications\n\nHigh\
    \ school diploma or GED equivalent required Must have resided in the United States\
    \ for at least three out of the last five years or worked for the U.S. in a foreign\
    \ country as either an employee or contractor in a federal or military capacity\
    \ for at least three of the last five yearsHaving your own Personally Owned Vehicle\
    \ (POV) is requiredPossess a demonstrated ability to analyze documents to extract\
    \ informationGood oral and written communication skillsHave hands-on familiarity\
    \ with a variety of computer applications,Must have a working knowledge of a variety\
    \ of computer software applications in word processing, spreadsheets, databases,\
    \ presentation software (MS Word, Excel, PowerPoint), and OutlookA valid driver’s\
    \ license is requiredTop Secret clearance required \n\n\nKnowledge, Skills, And\
    \ Abilities\n\nPossess a demonstrated ability to analyze documents to extract\
    \ informationGood oral and written communication skillsHave hands-on familiarity\
    \ with a variety of computer applications, including word processing, database,\
    \ spreadsheet, and telecommunications softwareMust be a team playerMust be able\
    \ to work independently and with USMS staff to interpret data rapidly and accurately\
    \ for proper execution in a records management databaseMust have a working knowledge\
    \ of a variety of computer software applications in word processing, spreadsheets,\
    \ databases, presentation software (MS Word, Excel, Access, PowerPoint), and OutlookAbility\
    \ to work independently on tasks be a self-starter and complete projects with\
    \ a team as they ariseAttention to detail and the ability to direct the work of\
    \ others efficiently and effectivelyAbility to consistently deliver high-quality\
    \ work under extreme pressureAbility to work shiftworkAbility to lift and move\
    \ boxes up to 25 pounds, including frequently utilizing hands, arms, and legs\
    \ for file placement and removalExperience with scanning software\n\n\nHow You’ll\
    \ Grow\n\nAt Chenega MIOS, our professional development plan focuses on helping\
    \ our team members at every level of their career to identify and use their strengths\
    \ to do their best work every day. From entry-level employees to senior leaders,\
    \ we believe there’s always room to learn.\n\nWe offer opportunities to help sharpen\
    \ skills in addition to hands-on experience in the global, fast-changing business\
    \ world. From on-the-job learning experiences to formal development programs,\
    \ our professionals have a variety of opportunities to continue to grow throughout\
    \ their careers.\n\nBenefits\n\nAt Chenega MIOS, we know that great people make\
    \ a great organization. We value our team members and offer them a broad range\
    \ of benefits.\n\nLearn more about what working at Chenega MIOS can mean for you.\n\
    \nChenega MIOS’s culture\n\nOur positive and supportive culture encourages our\
    \ team members to do their best work every day. We celebrate individuals by recognizing\
    \ their uniqueness and offering them the flexibility to make daily choices that\
    \ can help them be healthy, centered, confident, and aware. We offer well-being\
    \ programs and continuously look for new ways to maintain a culture where we excel\
    \ and lead healthy, happy lives.\n\nCorporate citizenship\n\nChenega MIOS is led\
    \ by a purpose to make an impact that matters. This purpose defines who we are\
    \ and extends to relationships with our clients, our team members, and our communities.\
    \ We believe that business has the power to inspire and transform. We focus on\
    \ education, giving, skill-based volunteerism, and leadership to help drive positive\
    \ social impact in our communities.\n\nLearn more about Chenega’s impact on the\
    \ world.\n\nChenega MIOS News- https://chenegamios.com/news/\n\nTips from your\
    \ Talent Acquisition team\n\nWe Want Job Seekers Exploring Opportunities At Chenega\
    \ MIOS To Feel Prepared And Confident. To Help You With Your Research, We Suggest\
    \ You Review The Following Links\n\nChenega MIOS web site - www.chenegamios.com\n\
    \nGlassdoor - https://www.glassdoor.com/Overview/Working-at-Chenega-MIOS-EI_IE369514.11,23.htm\n\
    \nLinkedIn - https://www.linkedin.com/company/1472684/\n\nFacebook - https://www.facebook.com/chenegamios/\n\
    \n#DICE\n\n#Chenega Defense & Aerospace Solutions, LLC"
  - "Qualifications\n Data Engineering, Data Modeling, and ETL (Extract Transform\
    \ Load) skillsData Warehousing and Data Analytics skillsExperience with data-related\
    \ tools and technologiesStrong problem-solving and analytical skillsExcellent\
    \ written and verbal communication skillsAbility to work independently and remotelyExperience\
    \ with cloud platforms (e.g., AWS, Azure) is a plusBachelor's degree in Computer\
    \ Science, Information Systems, or related field"
  - skills will be difficult. The more aligned skills they have, the better.Organizational
    Structure And Impact:Describe the function your group supports from an LOB perspective:Experienced
    ML engineer to work on universal forecasting models. Focus on ML forecasting,
    Python and Hadoop. Experience with Python, ARIMA, FB Prophet, Seasonal Naive,
    Gluon.Data Science Innovation (DSI) is a very unique application. It is truly
    ML-driven at its heart and our forecasting models originally looked singularly
    at cash balance forecasting. That has all changed as we have now incorporated
    approximately 100 additional financial metrics from our new DSI Metrics Farm.
    This allows future model executions to become a Universal Forecasting Model instead
    of being limited to just cash forecasting. It’s a very exciting application, especially
    since the models have been integrated within a Marketplace concept UI that allows
    Subscriber/Contributor functionality to make information and processing more personal
    and with greater extensibility across the enterprise. The application architecture
    is represented by OpenShift, Linux, Oracle, SQL Server, Hadoop, MongoDB, APIs,
    and a great deal of Python code.Describe the current initiatives that this resource
    will be impacting:Working toward implementation of Machine Learning Services.Team
    Background and Preferred Candidate History:Do you only want candidates with a
    similar background or would you like to see candidates with a diverse industry
    background?Diverse industry background, finance background preferred. Manager
    is more focused on the skillset.Describe the dynamic of your team and where this
    candidate will fit into the overall environment:This person will work with a variety
    of titles including application architects, web engineers, data engineers, data
    scientists, application system managers, system integrators, and Quality Engineers.Will
    work with various teams, but primarily working with one core team - approx 15
    - onshore and offshore resources.Candidate Technical and skills profile:Describe
    the role and the key responsibilities in order of which they will be doing daily:Machine
    Learning Engineer that work with Data Scientists in a SDLC environment into production.Interviews:Describe
    interview process (who will be involved, how many interviews, etc.):1 round -
    1 hour minimum, panel style
datasets:
- ShushantLLM/ai-job-embedding-finetuning
pipeline_tag: sentence-similarity
library_name: sentence-transformers
metrics:
- cosine_accuracy
model-index:
- name: SentenceTransformer based on sentence-transformers/paraphrase-multilingual-mpnet-base-v2
  results:
  - task:
      type: triplet
      name: Triplet
    dataset:
      name: ai job validation
      type: ai-job-validation
    metrics:
    - type: cosine_accuracy
      value: 0.9801980257034302
      name: Cosine Accuracy
  - task:
      type: triplet
      name: Triplet
    dataset:
      name: ai job test
      type: ai-job-test
    metrics:
    - type: cosine_accuracy
      value: 0.9607843160629272
      name: Cosine Accuracy
---

# SentenceTransformer based on sentence-transformers/paraphrase-multilingual-mpnet-base-v2

This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/paraphrase-multilingual-mpnet-base-v2](https://huggingface.co/sentence-transformers/paraphrase-multilingual-mpnet-base-v2) on the [ai-job-embedding-finetuning](https://huggingface.co/datasets/ShushantLLM/ai-job-embedding-finetuning) dataset. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.

## Model Details

### Model Description
- **Model Type:** Sentence Transformer
- **Base model:** [sentence-transformers/paraphrase-multilingual-mpnet-base-v2](https://huggingface.co/sentence-transformers/paraphrase-multilingual-mpnet-base-v2) <!-- at revision 4328cf26390c98c5e3c738b4460a05b95f4911f5 -->
- **Maximum Sequence Length:** 128 tokens
- **Output Dimensionality:** 768 dimensions
- **Similarity Function:** Cosine Similarity
- **Training Dataset:**
    - [ai-job-embedding-finetuning](https://huggingface.co/datasets/ShushantLLM/ai-job-embedding-finetuning)
<!-- - **Language:** Unknown -->
<!-- - **License:** Unknown -->

### Model Sources

- **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
- **Repository:** [Sentence Transformers on GitHub](https://github.com/huggingface/sentence-transformers)
- **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)

### Full Model Architecture

```
SentenceTransformer(
  (0): Transformer({'max_seq_length': 128, 'do_lower_case': False, 'architecture': 'XLMRobertaModel'})
  (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
)
```

## Usage

### Direct Usage (Sentence Transformers)

First install the Sentence Transformers library:

```bash
pip install -U sentence-transformers
```

Then you can load this model and run inference.
```python
from sentence_transformers import SentenceTransformer

# Download from the 🤗 Hub
model = SentenceTransformer("ShushantLLM/paraphrase-multilingual-mpnet-base-v2")
# Run inference
queries = [
    "Data organization, document analysis, records management",
]
documents = [
    'skills and build your career in a rapidly evolving business climate? Are you looking for a career where professional development is embedded in your employer’s core culture? If so, Chenega Military, Intelligence & Operations Support (MIOS) could be the place for you! Join our team of professionals who support large-scale government operations by leveraging cutting-edge technology and take your career to the next level!\n\nAs one of the newest Chenega companies, Chenega Defense & Aerospace Solutions (CDAS) was developed with the purpose of providing expert Engineering and Technical Support Services to federal customers.\n\nThe Data Analyst will analyze a large variety of documents to ensure proper placement in physical files, perform high-level scanning of master file documents to convert them into an electronic format, and provide meticulous organization and management of case files, including sorting and categorizing documents before scanning.\n\nResponsibilities\n\nWork within the Standard Operating Procedure for the organization of physical files containing documents of various types Establish or maintain physical files, including proper placement of documents as they are createdDisseminate significant amounts of information with attention to detail and accuracyPerform word processing tasksPerform data entry and metadata entry for electronic documentsReconcile inconsistenciesGather information and organize investigative packages, case files, or presentationsObtain additional information from other investigative agencies or databasesVerify information and files against the tracking systemMaintain internal status information on the disposition of designated information and filesDistribute and receive documentsAssist analyst or government official in obtaining or collecting all documents or information to complete case fileProvide administrative information and assistance concerning the case or files to other agencies or organizationsOther duties as assigned\n\n\nQualifications\n\nHigh school diploma or GED equivalent required Must have resided in the United States for at least three out of the last five years or worked for the U.S. in a foreign country as either an employee or contractor in a federal or military capacity for at least three of the last five yearsHaving your own Personally Owned Vehicle (POV) is requiredPossess a demonstrated ability to analyze documents to extract informationGood oral and written communication skillsHave hands-on familiarity with a variety of computer applications,Must have a working knowledge of a variety of computer software applications in word processing, spreadsheets, databases, presentation software (MS Word, Excel, PowerPoint), and OutlookA valid driver’s license is requiredTop Secret clearance required \n\n\nKnowledge, Skills, And Abilities\n\nPossess a demonstrated ability to analyze documents to extract informationGood oral and written communication skillsHave hands-on familiarity with a variety of computer applications, including word processing, database, spreadsheet, and telecommunications softwareMust be a team playerMust be able to work independently and with USMS staff to interpret data rapidly and accurately for proper execution in a records management databaseMust have a working knowledge of a variety of computer software applications in word processing, spreadsheets, databases, presentation software (MS Word, Excel, Access, PowerPoint), and OutlookAbility to work independently on tasks be a self-starter and complete projects with a team as they ariseAttention to detail and the ability to direct the work of others efficiently and effectivelyAbility to consistently deliver high-quality work under extreme pressureAbility to work shiftworkAbility to lift and move boxes up to 25 pounds, including frequently utilizing hands, arms, and legs for file placement and removalExperience with scanning software\n\n\nHow You’ll Grow\n\nAt Chenega MIOS, our professional development plan focuses on helping our team members at every level of their career to identify and use their strengths to do their best work every day. From entry-level employees to senior leaders, we believe there’s always room to learn.\n\nWe offer opportunities to help sharpen skills in addition to hands-on experience in the global, fast-changing business world. From on-the-job learning experiences to formal development programs, our professionals have a variety of opportunities to continue to grow throughout their careers.\n\nBenefits\n\nAt Chenega MIOS, we know that great people make a great organization. We value our team members and offer them a broad range of benefits.\n\nLearn more about what working at Chenega MIOS can mean for you.\n\nChenega MIOS’s culture\n\nOur positive and supportive culture encourages our team members to do their best work every day. We celebrate individuals by recognizing their uniqueness and offering them the flexibility to make daily choices that can help them be healthy, centered, confident, and aware. We offer well-being programs and continuously look for new ways to maintain a culture where we excel and lead healthy, happy lives.\n\nCorporate citizenship\n\nChenega MIOS is led by a purpose to make an impact that matters. This purpose defines who we are and extends to relationships with our clients, our team members, and our communities. We believe that business has the power to inspire and transform. We focus on education, giving, skill-based volunteerism, and leadership to help drive positive social impact in our communities.\n\nLearn more about Chenega’s impact on the world.\n\nChenega MIOS News- https://chenegamios.com/news/\n\nTips from your Talent Acquisition team\n\nWe Want Job Seekers Exploring Opportunities At Chenega MIOS To Feel Prepared And Confident. To Help You With Your Research, We Suggest You Review The Following Links\n\nChenega MIOS web site - www.chenegamios.com\n\nGlassdoor - https://www.glassdoor.com/Overview/Working-at-Chenega-MIOS-EI_IE369514.11,23.htm\n\nLinkedIn - https://www.linkedin.com/company/1472684/\n\nFacebook - https://www.facebook.com/chenegamios/\n\n#DICE\n\n#Chenega Defense & Aerospace Solutions, LLC',
    'skills will be difficult. The more aligned skills they have, the better.Organizational Structure And Impact:Describe the function your group supports from an LOB perspective:Experienced ML engineer to work on universal forecasting models. Focus on ML forecasting, Python and Hadoop. Experience with Python, ARIMA, FB Prophet, Seasonal Naive, Gluon.Data Science Innovation (DSI) is a very unique application. It is truly ML-driven at its heart and our forecasting models originally looked singularly at cash balance forecasting. That has all changed as we have now incorporated approximately 100 additional financial metrics from our new DSI Metrics Farm. This allows future model executions to become a Universal Forecasting Model instead of being limited to just cash forecasting. It’s a very exciting application, especially since the models have been integrated within a Marketplace concept UI that allows Subscriber/Contributor functionality to make information and processing more personal and with greater extensibility across the enterprise. The application architecture is represented by OpenShift, Linux, Oracle, SQL Server, Hadoop, MongoDB, APIs, and a great deal of Python code.Describe the current initiatives that this resource will be impacting:Working toward implementation of Machine Learning Services.Team Background and Preferred Candidate History:Do you only want candidates with a similar background or would you like to see candidates with a diverse industry background?Diverse industry background, finance background preferred. Manager is more focused on the skillset.Describe the dynamic of your team and where this candidate will fit into the overall environment:This person will work with a variety of titles including application architects, web engineers, data engineers, data scientists, application system managers, system integrators, and Quality Engineers.Will work with various teams, but primarily working with one core team - approx 15 - onshore and offshore resources.Candidate Technical and skills profile:Describe the role and the key responsibilities in order of which they will be doing daily:Machine Learning Engineer that work with Data Scientists in a SDLC environment into production.Interviews:Describe interview process (who will be involved, how many interviews, etc.):1 round - 1 hour minimum, panel style',
    "Qualifications\n Data Engineering, Data Modeling, and ETL (Extract Transform Load) skillsData Warehousing and Data Analytics skillsExperience with data-related tools and technologiesStrong problem-solving and analytical skillsExcellent written and verbal communication skillsAbility to work independently and remotelyExperience with cloud platforms (e.g., AWS, Azure) is a plusBachelor's degree in Computer Science, Information Systems, or related field",
]
query_embeddings = model.encode_query(queries)
document_embeddings = model.encode_document(documents)
print(query_embeddings.shape, document_embeddings.shape)
# [1, 768] [3, 768]

# Get the similarity scores for the embeddings
similarities = model.similarity(query_embeddings, document_embeddings)
print(similarities)
# tensor([[ 0.0065,  0.0405, -0.2204]])
```

<!--
### Direct Usage (Transformers)

<details><summary>Click to see the direct usage in Transformers</summary>

</details>
-->

<!--
### Downstream Usage (Sentence Transformers)

You can finetune this model on your own dataset.

<details><summary>Click to expand</summary>

</details>
-->

<!--
### Out-of-Scope Use

*List how the model may foreseeably be misused and address what users ought not to do with the model.*
-->

## Evaluation

### Metrics

#### Triplet

* Datasets: `ai-job-validation` and `ai-job-test`
* Evaluated with [<code>TripletEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.TripletEvaluator)

| Metric              | ai-job-validation | ai-job-test |
|:--------------------|:------------------|:------------|
| **cosine_accuracy** | **0.9802**        | **0.9608**  |

<!--
## Bias, Risks and Limitations

*What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
-->

<!--
### Recommendations

*What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
-->

## Training Details

### Training Dataset

#### ai-job-embedding-finetuning

* Dataset: [ai-job-embedding-finetuning](https://huggingface.co/datasets/ShushantLLM/ai-job-embedding-finetuning) at [1de228a](https://huggingface.co/datasets/ShushantLLM/ai-job-embedding-finetuning/tree/1de228a8cb18a24605027066b73f54957a2b9ce0)
* Size: 810 training samples
* Columns: <code>query</code>, <code>job_description_pos</code>, and <code>job_description_neg</code>
* Approximate statistics based on the first 810 samples:
  |         | query                                                                             | job_description_pos                                                                 | job_description_neg                                                                 |
  |:--------|:----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|
  | type    | string                                                                            | string                                                                              | string                                                                              |
  | details | <ul><li>min: 9 tokens</li><li>mean: 17.49 tokens</li><li>max: 43 tokens</li></ul> | <ul><li>min: 8 tokens</li><li>mean: 121.41 tokens</li><li>max: 128 tokens</li></ul> | <ul><li>min: 7 tokens</li><li>mean: 122.26 tokens</li><li>max: 128 tokens</li></ul> |
* Samples:
  | query                                                                                                      | job_description_pos                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                        | job_description_neg                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              |
  |:-----------------------------------------------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
  | <code>Senior Data Analyst, monitoring systems, dashboard development, statistical analysis</code>          | <code>experience where you can also make an impact on your community. While safety is a serious business, we are a supportive team that is optimizing the remote experience to create strong and fulfilling relationships even when we are physically apart. Our group of hard-working employees thrive in a positive and inclusive environment, where a bias towards action is rewarded.<br><br>We have raised over $380M in venture capital from investors including Tiger Global, Andreessen Horowitz, Matrix Partners, Meritech Capital Partners, and Initialized Capital. Now surpassing a $3B valuation, Flock is scaling intentionally and seeking the best and brightest to help us meet our goal of reducing crime in the United States by 25% in the next three years.<br><br>The Opportunity<br><br>As a Senior Data Analyst on the ML team, you will be responsible for extracting insights aggregated from various data sources, developing dashboards to identify trends and patterns that highlight model performance issues, performing analysis...</code> | <code>SKILLS and EXPERIENCE:3-5+ years of experience domain knowledge with either support of core Banking application experience, Mortgage Servicing or Loan Originations or personal or auto loans within Finance Industry environmentAble to interact with the VP or C-level Business Executives and higher to gather requirements and collaborate with IT; working effectively and independently as well as be collaborative team-oriented team player.Ideally supported Mortgage servicing systems such as Black Knight’s MSP, Sagent, Finastra’s Fusion Servicing Director, Interlinq Loan Servicing (ILS) or other loan servicing platform OR support of other core banking or originations platformSome experience with the following core technologies: T-SQL; SQL Server 2016 or higher; Visual Studio 2017 or higher; SQL Server Data Tools; Team Foundation ServerWorking knowledge of T-SQL programming and scripting, as well as optimization techniques· 3 years of experience with a strong focus on SQL Relational databases, ...</code>                         |
  | <code>advanced analytics, financial strategy, data visualization</code>                                    | <code>skills and business acumen to drive impactful results that inform strategic decisions.Commitment to iterative development, with a proven ability to engage and update stakeholders bi-weekly or as necessary, ensuring alignment, feedback incorporation, and transparency throughout the project lifecycle.Project ownership and development from inception to completion, encompassing tasks such as gathering detailed requirements, data preparation, model creation, result generation, and data visualization. Develop insights, methods or tools using various analytic methods such as causal-model approaches, predictive modeling, regressions, machine learning, time series analysis, etc.Handle large amounts of data from multiple and disparate sources, employing advanced Python and SQL techniques to ensure efficiency and accuracyUphold the highest standards of data integrity and security, aligning with both internal and external regulatory requirements and compliance protocols<br><br>Required Qualifications, C...</code>             | <code>experience Life at Visa.<br><br>Job Description<br><br>About the Team:<br><br>VISA is the leader in the payment industry and has been for a long time, but we are also quickly transitioning into a technology company that is fostering an environment for applying the newest technology to solve exciting problems in this area. For a payment system to work well, the risk techniques, performance, and scalability are critical. These techniques and systems benefit from big data, data mining, artificial intelligence, machine learning, cloud computing, & many other advance technologies. At VISA, we have all of these. If you want to be on the cutting edge of the payment space, learn fast, and make a big impact, then the Artificial Intelligence Platform team may be an ideal place for you!<br><br>Our team needs a Senior Data Engineer with proven knowledge of web application and web service development who will focus on creating new capabilities for the AI Platform while maturing our code base and development processes. You...</code> |
  | <code>Clinical Operations data analysis, eTMF, EDC implementation, advanced analytics visualization</code> | <code>requirements, and objectives for Clinical initiatives  Technical SME for system activities for the clinical system(s), enhancements, and integration projects. Coordinates support activities across vendor(s)  Systems include but are not limited to eTMF, EDC, CTMS and Analytics  Interfaces with external vendors at all levels to manage the relationship and ensure the proper delivery of services  Document Data Transfer Agreements for Data Exchange between BioNTech and Data Providers (CRO, Partner Organizations)  Document Data Transformation logic and interact with development team to convert business logic into technical details <br><br> What you have to offer: <br><br> Bachelor’s or higher degree in a scientific discipline (e.g., computer science/information systems, engineering, mathematics, natural sciences, medical, or biomedical science)  Extensive experience/knowledge of technologies and trends including Visualizations /Advanced Analytics  Outstanding analytical skills and result orientation  Ab...</code>       | <code>Requirements<br><br>Typically requires 13+ years of professional experience and 6+ years of diversified leadership, planning, communication, organization, and people motivation skills (or equivalent experience).<br><br>Critical Skills<br><br>12+ years of experience in a technology role; proven experience in a leadership role, preferably in a large, complex organization.8+ years Data Engineering, Emerging Technology, and Platform Design experience4+ years Leading large data / technical teams – Data Engineering, Solution Architects, and Business Intelligence Engineers, encouraging a culture of innovation, collaboration, and continuous improvement.Hands-on experience building and delivering Enterprise Data SolutionsExtensive market knowledge and experience with cutting edge Data, Analytics, Data Science, ML and AI technologiesExtensive professional experience with ETL, BI & Data AnalyticsExtensive professional experience with Big Data systems, data pipelines and data processingDeep expertise in Data Archit...</code>       |
* Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
  ```json
  {
      "scale": 20.0,
      "similarity_fct": "cos_sim",
      "gather_across_devices": false
  }
  ```

### Evaluation Dataset

#### ai-job-embedding-finetuning

* Dataset: [ai-job-embedding-finetuning](https://huggingface.co/datasets/ShushantLLM/ai-job-embedding-finetuning) at [1de228a](https://huggingface.co/datasets/ShushantLLM/ai-job-embedding-finetuning/tree/1de228a8cb18a24605027066b73f54957a2b9ce0)
* Size: 101 evaluation samples
* Columns: <code>query</code>, <code>job_description_pos</code>, and <code>job_description_neg</code>
* Approximate statistics based on the first 101 samples:
  |         | query                                                                              | job_description_pos                                                                  | job_description_neg                                                                  |
  |:--------|:-----------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|
  | type    | string                                                                             | string                                                                               | string                                                                               |
  | details | <ul><li>min: 10 tokens</li><li>mean: 17.83 tokens</li><li>max: 32 tokens</li></ul> | <ul><li>min: 18 tokens</li><li>mean: 122.03 tokens</li><li>max: 128 tokens</li></ul> | <ul><li>min: 17 tokens</li><li>mean: 119.95 tokens</li><li>max: 128 tokens</li></ul> |
* Samples:
  | query                                                                                                    | job_description_pos                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              | job_description_neg                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  |
  |:---------------------------------------------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
  | <code>Azure Data Factory, Databricks, Snowflake architecture</code>                                      | <code>Skills: SQL, PySpark, Databricks, Azure Synapse, Azure Data Factory.<br>Need hands-on coding<br>Requirements:1. Extensive knowledge of any of the big cloud services - Azure, AWS or GCP with practical implementation (like S3, ADLS, Airflow, ADF, Lamda, BigQuery, EC2, Fabric, Databricks or equivalent)2. Strong Hands-on experience in SQL and Python/PySpark programming knowledge. Should be able to write code during an interview with minimal syntax error.3. Strong foundational and architectural knowledge of any of the data warehouses - Snowflake, Redshift. Synapse etc.4. Should be able to drive and deliver projects with little or no guidance. Take ownership, become a self-learner, and have leadership qualities.</code>                                                                                                                                                                                                                                                                                                                         | <code>experience for yourself, and a better working world for all.<br><br>Data Analyst, Technology Consulting - Data & Analytics (Data Governance & Controls) - Financial Services Office (Manager) (Multiple Positions), Ernst & Young U.S. LLP, New York, NY. <br><br>Work with clients to transform the way they use and manage data by architecting data strategies, providing end-to-end solutions that focus on improving their data supply chain, reengineering processes, enhancing risk control, and enabling information intelligence by harnessing latest advanced technologies. Solve complex issues and drive growth across financial services. Define data and analytic strategies by performing assessments, recommending remediation strategies/solutions based on aggregated view of identified gaps, and designing/implementing future state data and analytics solutions. Manage and coach diverse teams of professionals with different backgrounds. Manage cross functional teams, to ensure project task and timeline accountability...</code> |
  | <code>Big Data Engineer, Spark, Hadoop, AWS GCP</code>                                                   | <code>Skills • Expertise and hands-on experience on Spark, and Hadoop echo system components – Must Have • Good and hand-on experience* of any of the Cloud (AWS/GCP) – Must Have • Good knowledge of HiveQL & SparkQL – Must Have Good knowledge of Shell script & Java/Scala/python – Good to Have • Good knowledge of SQL – Good to Have • Good knowledge of migration projects on Hadoop – Good to Have • Good Knowledge of one of the Workflow engines like Oozie, Autosys – Good to Have Good knowledge of Agile Development– Good to Have • Passionate about exploring new technologies – Good to Have • Automation approach – Good to Have <br>Thanks & RegardsShahrukh KhanEmail: shahrukh@zentekinfosoft.com</code>                                                                                                                                                                                                                                                                                                                                                    | <code>Requirements: We're looking for a candidate with exceptional proficiency in Google Sheets. This expertise should include manipulating, analyzing, and managing data within Google Sheets. The candidate should be outstanding at extracting business logic from existing reports and implementing it into new ones. Although a basic understanding of SQL for tasks related to data validation and metrics calculations is beneficial, the primary skill we are seeking is proficiency in Google Sheets. This role will involve working across various cross-functional teams, so strong communication skills are essential. The position requires a meticulous eye for detail, a commitment to delivering high-quality results, and above all, exceptional competency in Google Sheets<br><br>Google sheet knowledge is preferred.Strong Excel experience without Google will be considered.Data Validation and formulas to extract data are a mustBasic SQL knowledge is required.Strong communications skills are requiredInterview process...</code>       |
  | <code>Energy policy analysis, regulatory impact modeling, distributed energy resource management.</code> | <code>skills, modeling, energy data analysis, and critical thinking are required for a successful candidate. Knowledge of energy systems and distributed solar is required.<br><br>Reporting to the Senior Manager of Government Affairs, you will work across different teams to model data to inform policy advocacy. The ability to obtain data from multiple sources, including regulatory or legislative hearings, academic articles, and reports, are fundamental to the role.<br><br>A willingness to perform under deadlines and collaborate within an organization is required. Honesty, accountability, and integrity are a must.<br><br>Energy Policy & Data Analyst Responsibilities<br><br>Support Government Affairs team members with energy policy recommendations based on data modelingEvaluate relevant regulatory or legislative filings and model the impacts to Sunnova’s customers and businessAnalyze program proposals (grid services, incentives, net energy metering, fixed charges) and develop recommendations that align with Sunnova’s ...</code> | <code>QualificationsData Engineering, Data Modeling, and ETL (Extract Transform Load) skillsMonitor and support data pipelines and ETL workflowsData Warehousing and Data Analytics skillsExperience with Azure cloud services and toolsStrong problem-solving and analytical skillsProficiency in SQL and other programming languagesExperience with data integration and data migrationExcellent communication and collaboration skillsBachelor's degree in Computer Science, Engineering, or related field<br>Enterprise Required SkillsPython, Big data, Data warehouse, ETL, Development, azure, Azure Data Factory, Azure Databricks, Azure SQL Server, Snowflake, data pipelines<br>Top Skills Details1. 3+ years with ETL Development with Azure stack (Azure Data Factory, Azure Databricks, Azure Blob, Azure SQL).  2. 3+ years with Spark, SQL, and Python. This will show up with working with large sets of data in an enterprise environment.  3. Looking for Proactive individuals who have completed projects from start to complet...</code>       |
* Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
  ```json
  {
      "scale": 20.0,
      "similarity_fct": "cos_sim",
      "gather_across_devices": false
  }
  ```

### Training Hyperparameters
#### Non-Default Hyperparameters

- `eval_strategy`: steps
- `per_device_train_batch_size`: 16
- `per_device_eval_batch_size`: 16
- `learning_rate`: 2e-05
- `num_train_epochs`: 5
- `warmup_ratio`: 0.1
- `batch_sampler`: no_duplicates

#### All Hyperparameters
<details><summary>Click to expand</summary>

- `overwrite_output_dir`: False
- `do_predict`: False
- `eval_strategy`: steps
- `prediction_loss_only`: True
- `per_device_train_batch_size`: 16
- `per_device_eval_batch_size`: 16
- `per_gpu_train_batch_size`: None
- `per_gpu_eval_batch_size`: None
- `gradient_accumulation_steps`: 1
- `eval_accumulation_steps`: None
- `torch_empty_cache_steps`: None
- `learning_rate`: 2e-05
- `weight_decay`: 0.0
- `adam_beta1`: 0.9
- `adam_beta2`: 0.999
- `adam_epsilon`: 1e-08
- `max_grad_norm`: 1.0
- `num_train_epochs`: 5
- `max_steps`: -1
- `lr_scheduler_type`: linear
- `lr_scheduler_kwargs`: {}
- `warmup_ratio`: 0.1
- `warmup_steps`: 0
- `log_level`: passive
- `log_level_replica`: warning
- `log_on_each_node`: True
- `logging_nan_inf_filter`: True
- `save_safetensors`: True
- `save_on_each_node`: False
- `save_only_model`: False
- `restore_callback_states_from_checkpoint`: False
- `no_cuda`: False
- `use_cpu`: False
- `use_mps_device`: False
- `seed`: 42
- `data_seed`: None
- `jit_mode_eval`: False
- `bf16`: False
- `fp16`: False
- `fp16_opt_level`: O1
- `half_precision_backend`: auto
- `bf16_full_eval`: False
- `fp16_full_eval`: False
- `tf32`: None
- `local_rank`: 0
- `ddp_backend`: None
- `tpu_num_cores`: None
- `tpu_metrics_debug`: False
- `debug`: []
- `dataloader_drop_last`: False
- `dataloader_num_workers`: 0
- `dataloader_prefetch_factor`: None
- `past_index`: -1
- `disable_tqdm`: False
- `remove_unused_columns`: True
- `label_names`: None
- `load_best_model_at_end`: False
- `ignore_data_skip`: False
- `fsdp`: []
- `fsdp_min_num_params`: 0
- `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
- `fsdp_transformer_layer_cls_to_wrap`: None
- `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
- `parallelism_config`: None
- `deepspeed`: None
- `label_smoothing_factor`: 0.0
- `optim`: adamw_torch_fused
- `optim_args`: None
- `adafactor`: False
- `group_by_length`: False
- `length_column_name`: length
- `project`: huggingface
- `trackio_space_id`: trackio
- `ddp_find_unused_parameters`: None
- `ddp_bucket_cap_mb`: None
- `ddp_broadcast_buffers`: False
- `dataloader_pin_memory`: True
- `dataloader_persistent_workers`: False
- `skip_memory_metrics`: True
- `use_legacy_prediction_loop`: False
- `push_to_hub`: False
- `resume_from_checkpoint`: None
- `hub_model_id`: None
- `hub_strategy`: every_save
- `hub_private_repo`: None
- `hub_always_push`: False
- `hub_revision`: None
- `gradient_checkpointing`: False
- `gradient_checkpointing_kwargs`: None
- `include_inputs_for_metrics`: False
- `include_for_metrics`: []
- `eval_do_concat_batches`: True
- `fp16_backend`: auto
- `push_to_hub_model_id`: None
- `push_to_hub_organization`: None
- `mp_parameters`: 
- `auto_find_batch_size`: False
- `full_determinism`: False
- `torchdynamo`: None
- `ray_scope`: last
- `ddp_timeout`: 1800
- `torch_compile`: False
- `torch_compile_backend`: None
- `torch_compile_mode`: None
- `include_tokens_per_second`: False
- `include_num_input_tokens_seen`: no
- `neftune_noise_alpha`: None
- `optim_target_modules`: None
- `batch_eval_metrics`: False
- `eval_on_start`: False
- `use_liger_kernel`: False
- `liger_kernel_config`: None
- `eval_use_gather_object`: False
- `average_tokens_across_devices`: True
- `prompts`: None
- `batch_sampler`: no_duplicates
- `multi_dataset_batch_sampler`: proportional
- `router_mapping`: {}
- `learning_rate_mapping`: {}

</details>

### Training Logs
| Epoch  | Step | Training Loss | Validation Loss | ai-job-validation_cosine_accuracy | ai-job-test_cosine_accuracy |
|:------:|:----:|:-------------:|:---------------:|:---------------------------------:|:---------------------------:|
| -1     | -1   | -             | -               | 0.8416                            | -                           |
| 1.9608 | 100  | 1.2457        | 1.3444          | 0.9802                            | -                           |
| 3.9216 | 200  | 0.3222        | 1.3620          | 0.9802                            | -                           |
| -1     | -1   | -             | -               | 0.9802                            | 0.9608                      |


### Framework Versions
- Python: 3.12.12
- Sentence Transformers: 5.1.2
- Transformers: 4.57.1
- PyTorch: 2.8.0+cu126
- Accelerate: 1.11.0
- Datasets: 4.0.0
- Tokenizers: 0.22.1

## Citation

### BibTeX

#### Sentence Transformers
```bibtex
@inproceedings{reimers-2019-sentence-bert,
    title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
    author = "Reimers, Nils and Gurevych, Iryna",
    booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
    month = "11",
    year = "2019",
    publisher = "Association for Computational Linguistics",
    url = "https://arxiv.org/abs/1908.10084",
}
```

#### MultipleNegativesRankingLoss
```bibtex
@misc{henderson2017efficient,
    title={Efficient Natural Language Response Suggestion for Smart Reply},
    author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil},
    year={2017},
    eprint={1705.00652},
    archivePrefix={arXiv},
    primaryClass={cs.CL}
}
```

<!--
## Glossary

*Clearly define terms in order to be accessible across audiences.*
-->

<!--
## Model Card Authors

*Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
-->

<!--
## Model Card Contact

*Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
-->