metadata
tags:
- sentence-transformers
- sentence-similarity
- feature-extraction
- dense
- generated_from_trainer
- dataset_size:810
- loss:MultipleNegativesRankingLoss
base_model: sentence-transformers/paraphrase-multilingual-mpnet-base-v2
widget:
- source_sentence: CBRN defense, predictive analytics, natural language understanding
sentences:
- >-
experience with speech interfaces Lead and evaluate changing dialog
evaluation conventions, test tooling developments, and pilot processes
to support expansion to new data areas Continuously evaluate workflow
tools and processes and offer solutions to ensure they are efficient,
high quality, and scalable Provide expert support for a large and
growing team of data analysts Provide support for ongoing and new data
collection efforts as a subject matter expert on conventions and use of
the data Conduct research studies to understand speech and
customer-Alexa interactions Assist scientists, program and product
managers, and other stakeholders in defining and validating customer
experience metrics
We are open to hiring candidates to work out of one of the following
locations:
Boston, MA, USA | Seattle, WA, USA
Basic Qualifications
3+ years of data querying languages (e.g. SQL), scripting languages (e.g. Python) or statistical/mathematical software (e.g. R, SAS, Matlab, etc.) experience 2+ years of data scientist experience Bachelor's degree Experience applying theoretical models in an applied environment
Preferred Qualifications
Experience in Python, Perl, or another scripting language Experience in a ML or data scientist role with a large technology company Master's degree in a quantitative field such as statistics, mathematics, data science, business analytics, economics, finance, engineering, or computer science
Amazon is committed to a diverse and inclusive workplace. Amazon is
Our compensation reflects the cost of labor across several US geographic
markets. The base pay for this position ranges from $111,600/year in our
lowest geographic market up to $212,800/year in our highest geographic
market. Pay is based on a number of factors including market location
and may vary depending on job-related knowledge, skills, and experience.
Amazon is a total compensation company. Dependent on the position
offered, equity, sign-on payments, and other forms of compensation may
be provided as part of a total compensation package, in addition to a
full range of medical, financial, and/or other benefits. For more
information, please visit
https://www.aboutamazon.com/workplace/employee-benefits. This position
will remain posted until filled. Applicants should apply via our
internal or external career site.
Company - Amazon.com Services LLC
Job ID: A2610750
- >-
Skills: Your Expertise:
5+ years in industry experience and a degree (Masters or PhD is a plus)
in a quantitative field (e.g., Statistics, Econometrics, Computer
Science, Engineering, Mathematics, Data Science, Operations
Research).Expert communication and collaboration skills with the ability
to work effectively with internal teams in a cross-cultural and
cross-functional environment. Ability to conduct rigorous analysis and
communicate conclusions to both technical and non-technical
audiencesExperience partnering with internal teams to drive action and
providing expertise and direction on analytics, data science,
experimental design, and measurementExperience in analysis of A|B
experiments and statistical data analysisExperience designing and
building metrics, from conception to building prototypes with data
pipelinesStrong knowledge in at least one programming language (Python
or R) and in SQLAbility to drive data strategies, with a central source
of truth to impact business decisionsKnowledge and experience in
insurance industry - a plusKnowledge and experience in customer
experience measurement - a plus
Keywords:Education: Minimum: BS/BA in CS or related field (or
self-taught/ equivalent work experience) Preferred: MS/MA in CS or
related field
- >-
requirements of the program or company.
Working across the globe, V2X builds smart solutions designed to integrate physical and digital infrastructure from base to battlefield. We bring 120 years of successful mission support to improve security, streamline logistics, and enhance readiness. Aligned around a shared purpose, our $3.9B company and 16,000 people work alongside our clients, here and abroad, to tackle their most complex challenges with integrity, respect, responsibility, and professionalism.
At V2X, we are making a difference by delivering decision support tools
critical for the protection of our forces when threatened by both
physical and Chemical, Biological, Radiological, or Nuclear (CBRN)
threats.
We are expanding in data science to provide the best information
possible utilizing the latest techniques in Machine Learning (including
Deep Learning, Neural network). We are on the forefront of CBRN defense
and we are looking for talented Data Scientists that have applied
experience in the fields of artificial intelligence, machine learning
and/or natural language processing to join our team. Our data scientists
work closely everyday with project managers, subject matter experts and
software engineers to work on challenges in machine intelligence, data
mining, and machine learning, and work together with agility to build
capabilities that impress our customers. We partner and collaborate with
universities to being best minds together.
Data scientists will have opportunities to work on projects with highest
priority to our business. Vital to success in this role is the ability
to determine, define and deploy predictive / prescriptive analytic
solutions to identify and perform root cause analysis on adverse trends,
by choosing best fit methods, defining algorithms, and validating and
deploying models to achieve results.
Responsibilities
Major Job Activities:
Partner with our development teams to solve problems and identify trends and opportunities to leverage data from multiple sources. Collaborate across multiple teams. Passionate about working with large and complex unstructured and structured data sets. Strong communication and interpersonal skills. You should be able to work across functions and effectively present, recommend and communicate a position by demonstrating its value and tradeoffs. Comfortable conducting design, algorithm, and code reviews. Able to self-direct and succeed with minimal guidance.
Material & Equipment Directly Used:
Computer, Phone, and basic office materials.
Working Environment:
Function in an office environment in a stationary position approximately 50 percent of the time or more. Must be able to operate standard office equipment, such as a computer, copy machine, and printer.
Qualifications
Education / Certifications:
Bachelor’s degree in a computer, engineering, or quantitative discipline (e.g., statistics, operations research, bioinformatics, economics, computational biology, computer science, mathematics, physics, electrical engineering, industrial engineering). Master's or Ph.D. in a quantitative discipline preferred.
Clearance Requirement:
Must have or be able to obtain an active U.S. DoD Secret (or higher)
level Security Clearance.
Experience / Skills:
5+ years of relevant work experience in data analysis or related field. (e.g., statistician, data analyst, data scientist). Programming experience in one or more of the following: R, MATLAB, C, C++, Java, Python, Scala Experience in Natural Language Understanding, Computer Vision, Machine Learning, Algorithmic Foundations of Optimization, Data Mining or Machine Intelligence (Artificial Intelligence). Experience with statistical software (e.g., R, Octave, Julia, MATLAB, pandas) and database languages (e.g., SQL). Experience with machine learning related open source libraries including, but not limited to: Hadoop, Spark, SciKit-Learn, TensorFlow, etc. Contribution to research communities and/or efforts, including publishing papers at conferences.
We are committed to an inclusive and diverse workplace that values and
supports the contributions of each individual. This commitment along
with our common Vision and Values of Integrity, Respect, and
Responsibility, allows us to leverage differences, encourage innovation
and expand our success in the global marketplace. V2X is an Equal
Opportunity /Affirmative Action Employer. All qualified applicants will
receive consideration for employment without regard to race, color,
religion, age, sex, national origin, protected veteran status or status
as an individual with a disability.
- source_sentence: >-
Senior Data Analyst Pricing, data product automation, pricing strategy
analysis
sentences:
- >-
Skills You BringBachelor’s or Master’s Degree in a technology related
field (e.g. Engineering, Computer Science, etc.) required with 6+ years
of experienceInformatica Power CenterGood experience with ETL
technologiesSnaplogicStrong SQLProven data analysis skillsStrong data
modeling skills doing either Dimensional or Data Vault modelsBasic AWS
Experience Proven ability to deal with ambiguity and work in fast paced
environmentExcellent interpersonal and communication skillsExcellent
collaboration skills to work with multiple teams in the organization
- >-
experience, an annualized transactional volume of $140 billion in 2023,
and approximately 3,200 employees located in 12+ countries, Paysafe
connects businesses and consumers across 260 payment types in over 40
currencies around the world. Delivered through an integrated platform,
Paysafe solutions are geared toward mobile-initiated transactions,
real-time analytics and the convergence between brick-and-mortar and
online payments. Further information is available at www.paysafe.com.
Are you ready to make an impact? Join our team that is inspired by a
unified vision and propelled by passion.
Position Summary
We are looking for a dynamic and flexible, Senior Data Analyst, Pricing
to support our global Sales and Product organizations with strategic
planning, analysis, and commercial pricing efforts . As a Senior Data
Analyst , you will be at the frontier of building our Pricing function
to drive growth through data and AI-enabled capabilities. This
opportunity is high visibility for someone hungry to drive the upward
trajectory of our business and be able to contribute to their efforts in
the role in our success.
You will partner with Product Managers to understand their commercial
needs, then prioritize and work with a cross-functional team to deliver
pricing strategies and analytics-based solutions to solve and execute
them. Business outcomes will include sustainable growth in both revenues
and gross profit.
This role is based in Jacksonville, Florida and offers a flexible hybrid
work environment with 3 days in the office and 2 days working remote
during the work week.
Responsibilities
Build data products that power the automation and effectiveness of our pricing function, driving better quality revenues from merchants and consumers. Partner closely with pricing stakeholders (e.g., Product, Sales, Marketing) to turn raw data into actionable insights. Help ask the right questions and find the answers. Dive into complex pricing and behavioral data sets, spot trends and make interpretations. Utilize modelling and data-mining skills to find new insights and opportunities. Turn findings into plans for new data products or visions for new merchant features. Partner across merchant Product, Sales, Marketing, Development and Finance to build alignment, engagement and excitement for new products, features and initiatives. Ensure data quality and integrity by following and enforcing data governance policies, including alignment on data language.
Qualifications
Bachelor’s degree in a related field of study (Computer Science, Statistics, Mathematics, Engineering, etc.) required. 5+ years of experience of in-depth data analysis role, required; preferably in pricing context with B2B & B2C in a digital environment. Proven ability to visualize data intuitively, cleanly and clearly in order to make important insights simplified. Experience across large and complex datasets, including customer behavior, and transactional data. Advanced in SQL and in Python, preferred. Experience structuring and analyzing A/B tests, elasticities and interdependencies, preferred. Excellent communication and presentation skills, with the ability to explain complex data insights to non-technical audiences.
Life at Paysafe:
One network. One partnership. At Paysafe, this is not only our business
model; this is our mindset when it comes to our team. Being a part of
Paysafe means you’ll be one of over 3,200 members of a world-class team
that drives our business to new heights every day and where we are
committed to your personal and professional growth.
Our culture values humility, high trust & autonomy, a desire for
excellence and meeting commitments, strong team cohesion, a sense of
urgency, a desire to learn, pragmatically pushing boundaries, and
accomplishing goals that have a direct business impact.
Paysafe provides equal employment opportunities to all employees, and
applicants for employment, and prohibits discrimination of any type
concerning ethnicity, religion, age, sex, national origin, disability
status, sexual orientation, gender identity or expression, or any other
protected characteristics. This policy applies to all terms and
conditions of recruitment and employment. If you need any reasonable
adjustments, please let us know. We will be happy to help and look
forward to hearing from you.
- "Experience : 10 yearsLocation : RemoteDuration: Full TimeJob DetailsData Warehouse, ETL, Advanced SQL,Data Profiling, Source to Target Mapping,Business Requirement Document, FRS, Healthcare.Should be able to navigate the code - developer background\n\uFEFFThanks & Regard's\nMohd FurquanLead Technical RecruiterE-mail: [email protected] No: +1 925 313 8949LinkedIn-ID :linkedin.com/in/mohd-furquan-94237816aVisit us: www.msrcosmos.com"
- source_sentence: >-
CPG data analysis, Nielsen IRI expertise, Power Query dashboard
development
sentences:
- >-
Skills :
a) Azure Data Factory – Min 3 years of project experiencea. Design of
pipelinesb. Use of project with On-prem to Cloud Data Migrationc.
Understanding of ETLd. Change Data Capture from Multiple Sourcese. Job
Schedulingb) Azure Data Lake – Min 3 years of project experiencea. All
steps from design to deliverb. Understanding of different Zones and
design principalc) Data Modeling experience Min 5 Yearsa. Data
Mart/Warehouseb. Columnar Data design and modelingd) Reporting using
PowerBI Min 3 yearsa. Analytical Reportingb. Business Domain Modeling
and data dictionary
Interested please apply to the job, looking only for W2 candidates.
- >-
experienced and highly skilled Sr Data Engineer to join us. This role
requires a seasoned professional with a deep understanding of automated
data pipelines, cloud infrastructure, databases, and workflow engines.
The ideal candidate will have a minimum of 5 years of technical lead
experience in the medical device field and at least 7 years of
experience in data engineering. Proficiency in Python and a proven track
record of leading projects to completion are essential.
Primary Duties
Design, develop, and manage robust, secure, scalable, highly available,
and dynamic solutions to drive business objectives. Lead the
architecture and implementation of advanced cloud-based data engineering
solutions, leveraging AWS technologies and best practices. Manage and
optimize data pipelines, ensuring timely and accurate data availability
for analytics and machine learning applications. Oversee the
administration and performance tuning of databases and workflow engines.
Collaborate with cross-functional teams (e.g., product management, IT,
software engineering) to define data requirements, integrate systems,
and implement data governance and security policies. Mentor junior data
engineers and oversee the team's development efforts, promoting best
practices in coding, architecture, and data management. Stay abreast of
emerging technologies and trends in data engineering, cloud services,
and the medical device industry to drive innovation and competitive
advantage.
Knowledge, Experience & Skills
Degree in Computer Science, Engineering, Information Systems, or a
related field. Requiring a minimum of Bachelor’s degree +7yrs of
experience or a Master’s degree +5yrs of experience. Minimum of 7 years
of experience in data engineering, with expertise in developing and
managing automated data pipelines, AWS cloud infrastructure, databases,
and workflow engines. Certifications in AWS and data engineering
preferred. Experience with machine learning algorithms and data modeling
techniques. At least 5 years of experience in the medical device IVD
industry, with a strong understanding of FDA regulatory standards and
compliance requirements. Expert proficiency in Python programming and
software engineering principles. Demonstrated experience with AWS
services (e.g., EC2, RDS, S3, Lambda, Glue, Redshift, Athena, EMR) and
data pipeline tools (e.g., Apache Airflow, Luigi, etc). Strong knowledge
of database management (Postgres and Snowflake), SQL, and NoSQL
databases. Adept at queries, report writing and presenting findings
Experienced in developing and maintaining ETL pipelines in a cloud
environmentExperienced in Unit Testing preferred Strong analytical
skills with the ability to organize, analyze, and disseminate
information with attention to detail and accuracy Excellent
communication and task management skills. Comfort working in a dynamic,
fast-paced, research-oriented group with several ongoing concurrent
projectsFull fluency (verbal and written) of the English language is a
must.
The estimated salary range for this role based in California is between
$148,700 and $178,400 annually. This role is eligible to receive a
variable annual bonus based on company, team, and individual performance
per bioMerieux’s bonus program. This range may differ from ranges
offered for similar positions elsewhere in the country given differences
in cost of living. Actual compensation within this range is determined
based on the successful candidate’s experience and will be presented in
writing at the time of the offer.
In addition, bioMérieux offers a competitive Total Rewards package that
may include:
A choice of medical (including prescription), dental, and vision plans
providing nationwide coverage and telemedicine optionsCompany-Provided
Life and Accidental Death InsuranceShort and Long-Term Disability
InsuranceRetirement Plan including a generous non-discretionary employer
contribution and employer match. Adoption AssistanceWellness
ProgramsEmployee Assistance ProgramCommuter BenefitsVarious voluntary
benefit offeringsDiscount programsParental leaves
BioFire Diagnostics, LLC. is an Equal Opportunity/Affirmative Action
Employer. All qualified applicants will receive consideration for
employment without regard to race, color, religion, sex, sexual
orientation, gender identity, national origin, age, protected veteran or
disabled status, or genetic information.
Please be advised that the receipt of satisfactory responses to
reference requests and the provision of satisfactory proof of an
applicant’s identity and legal authorization to work in the United
States are required of all new hires. Any misrepresentation,
falsification, or material omission may result in the failure to receive
an offer, the retraction of an offer, or if already hired, dismissal. If
you are a qualified individual with a disability, you may request a
reasonable accommodation in BioFire Diagnostics’ application process by
contacting us via telephone at (385) 770-1132, by email at [email
protected], or by dialing 711 for access to Telecommunications Relay
Services (TRS).
- >-
requirements into analytical frameworks.Dashboard Development: Design
and maintain dashboards using Power Query in Excel, good in analytics in
generating metrics & measures and ensuring accurate and real-time data
representation.
Required QualificationsProfessional Experience: 3-6 years as a business
analyst, with mandatory experience in the CPG sector and should have
worked on brand dataTechnical Proficiency: Advanced skills in Excel and
Power Query;Communication Skills: Exceptional ability to communicate
complex data insights to non-technical stakeholders.Location: Position
based in Springdale. Preferred AttributesProven experience in
data-driven decision-making processes.Ability to handle multiple
projects simultaneously, with a focus on deadlines and results.
- source_sentence: ETL Pipelines, Apache Spark, AirFlow
sentences:
- >-
Qualifications
- Currently enrolled in a Bachelor’s or Master’s degree in Software Development, Computer Science, Computer Engineering, or a related technical discipline
- Must obtain work authorization in country of employment at the time of
hire, and maintain ongoing work authorization during employment.
Preferred Qualifications:
- Fluency in SQL or other programming languages (Python, R etc) for data
manipulation
- Ability to thrive in a fast paced work environment
- Ability to drive projects to completion with minimal guidance
- Ability to communicate the results of analyses in a clear and
effective manner
TikTok is committed to creating an inclusive space where employees are
valued for their skills, experiences, and unique perspectives. Our
platform connects people from across the globe and so does our
workplace. At TikTok, our mission is to inspire creativity and bring
joy. To achieve that goal, we are committed to celebrating our diverse
voices and to creating an environment that reflects the many communities
we reach. We are passionate about this and hope you are too.
TikTok is committed to providing reasonable accommodations in our
recruitment processes for candidates with disabilities, pregnancy,
sincerely held religious beliefs or other reasons protected by
applicable laws. If you need assistance or a reasonable accommodation,
please reach out to us at https://shorturl.at/cdpT2
By submitting an application for this role, you accept and agree to our
global applicant privacy policy, which may be accessed here:
https://careers.tiktok.com/legal/privacy.
Job Information:
【For Pay Transparency】Compensation Description (annually) The base
salary range for this position in the selected city is $45 -
$45annually. We cover 100% premium coverage for Full-Time intern medical
insurance after 90 days from the date of hire. Medical coverage only, no
dental or vision coverage.Our time off and leave plans are: Paid
holidays and paid sick leave. The sick leave entitlement is based on the
time you join.We also provide mental and emotional health benefits
through our Employee Assistance Program and provide reimbursements for
your mobile phone expense. The Company reserves the right to modify or
change these benefits programs at any time, with or without notice.
- >-
Experience as a Product Data Analyst at TGG:Achieving business results
as a client facing consultant for our clients in various types of
engagements within a variety of industries.Delivering high quality work
to our clients within our technology service line. Being part of a
collaborative, values-based firm that has a reputation for great work
and satisfied clients.Working with senior IT leaders to communicate
strategic goals to their organization, including leading client and
internal development teams on best practices.
What You Will Work On:Analyze large datasets to identify patterns,
trends, and opportunities for product optimization.Develop and maintain
dashboards and reports to track key performance metrics.Collaborate with
product managers, marketers, and engineers to ideate, prioritize, and
implement data-driven initiatives.Conduct A/B testing and other
statistical analyses to evaluate the effectiveness of product
changes.Communicate findings and recommendations to stakeholders through
clear and concise presentations.Contribute analytical insights to inform
product vision and deliver value.
Who Will You Work With:Client stakeholders ranging from individual
contributors to senior executives.A collaborative team of consultants
that deliver outstanding client service.TGG partners, principals,
account leaders, managers, and staff supporting you to excel within
client projects and to achieve your professional development goals.
Examples of What You Bring to the Table:You have strong analysis
capabilities and thrive on working collaboratively to deliver successful
results for clients. You have experience with these
technologies:Proficiency in SQL and Python for data extraction,
manipulation, and analysis.Strong understanding of statistical concepts
and techniques.Intermediate experience with Tableau, Power BI, Adobe
Analytics, or similar BI tools.Ability to analyze requirements, design,
implement, debug, and deploy Cloud Platform services and components.At
least basic exposure to data science and machine learning
methods.Familiarity with source control best practices: Define,
Setup/Configure, Deploy and Maintain source code (e.g. GIT, VisualSafe
Source).Ability to develop and schedule processes to extract, transform,
and store data from these systems: SQL databases, Azure cloud services,
Google cloud service, Snowflake.4-8 years of relevant
experience.Bachelor’s degree in Computer Science, Statistics, Economics,
Mathematics, or a related field; or equivalent combination of education,
training, and experience.Analytical Product Mindset: Ability to approach
problems analytically and derive actionable insights from complex
datasets, while remaining focused on providing value to customers
Strategic Thinking: Demonstrated ability to translate data findings into
strategic, achievable recommendations to drive business
outcomes.Communication Skills: Excellent verbal and written
communication skills.Ability to effectively convey technical concepts
from technical to non-technical stakeholders and vice-versa.Team Player:
Proven track record of collaborating effectively with cross-functional
teams in a fast-paced environment.Adaptability: Have consistently
demonstrated the ability to bring structure to complex, unstructured
environments.Familiarity with Agile development methodologies.Ability to
adapt to changing priorities to thrive in dynamic work environments.
Salary and Benefits:Nothing is more important to us than the well-being
of our team. That is why we are proud to offer a full suite of
competitive health benefits along with additional benefits such as:
flexible PTO, a professional development stipend and work from home
stipend, volunteer opportunities, and team social activities.
Salaries vary and are dependent on considerations such as: experience
and specific skills/certifications. The base plus target bonus total
compensation range for this role is $95,000 - $125,000. Additional
compensation beyond this range is available as a result of leadership
and business development opportunities. Salary details are discussed
openly during the hiring process.
Work Environment:TGG is headquartered in Portland, Oregon, and has team
members living in various locations across the United States. Our
consultants must have the ability to travel and to work remotely or
onsite. Each engagement has unique conditions, and we work
collaboratively to meet both our client and team's needs regarding
onsite and travel requirements.
Why The Gunter Group:TGG was created to be different, to be relational,
to be insightful, and to maximize potential for our consultants, our
clients, and our community. We listen first so we can learn, analyze,
and deliver meaningful solutions for our clients. Our compass points
towards our people and our “Non-Negotiables” always. Our driven
employees make us who we are — a talented team of leaders with deep and
diverse professional experience.If you think this role is the right fit,
please submit your resume and cover letter so we can learn more about
you.
The Gunter Group LLC is
- >-
Requirements & Day-to-Day: Design, develop, and support scalable data
processing pipelines using Apache Spark and Java/Scala. Lead a talented
team and make a significant impact on our data engineering capabilities.
Implement and manage workflow orchestration with AirFlow for efficient
data processing. Proficiently use SQL for querying and data manipulation
tasks. Collaborate with cross-functional teams to gather requirements
and ensure alignment with data engineering solutions. Essential
Criteria: a bachelor’s degree in computer science or another relevant
discipline, and a minimum of five years of relevant experience in data
engineering. Solid experience with Apache Spark for large-scale data
processing. Proficiency in Java or Scala programming languages. Strong
knowledge of AirFlow for workflow orchestration. Proficient in SQL for
data querying and manipulation.
- source_sentence: Data organization, document analysis, records management
sentences:
- >-
skills and build your career in a rapidly evolving business climate? Are
you looking for a career where professional development is embedded in
your employer’s core culture? If so, Chenega Military, Intelligence &
Operations Support (MIOS) could be the place for you! Join our team of
professionals who support large-scale government operations by
leveraging cutting-edge technology and take your career to the next
level!
As one of the newest Chenega companies, Chenega Defense & Aerospace
Solutions (CDAS) was developed with the purpose of providing expert
Engineering and Technical Support Services to federal customers.
The Data Analyst will analyze a large variety of documents to ensure
proper placement in physical files, perform high-level scanning of
master file documents to convert them into an electronic format, and
provide meticulous organization and management of case files, including
sorting and categorizing documents before scanning.
Responsibilities
Work within the Standard Operating Procedure for the organization of
physical files containing documents of various types Establish or
maintain physical files, including proper placement of documents as they
are createdDisseminate significant amounts of information with attention
to detail and accuracyPerform word processing tasksPerform data entry
and metadata entry for electronic documentsReconcile
inconsistenciesGather information and organize investigative packages,
case files, or presentationsObtain additional information from other
investigative agencies or databasesVerify information and files against
the tracking systemMaintain internal status information on the
disposition of designated information and filesDistribute and receive
documentsAssist analyst or government official in obtaining or
collecting all documents or information to complete case fileProvide
administrative information and assistance concerning the case or files
to other agencies or organizationsOther duties as assigned
Qualifications
High school diploma or GED equivalent required Must have resided in the
United States for at least three out of the last five years or worked
for the U.S. in a foreign country as either an employee or contractor in
a federal or military capacity for at least three of the last five
yearsHaving your own Personally Owned Vehicle (POV) is requiredPossess a
demonstrated ability to analyze documents to extract informationGood
oral and written communication skillsHave hands-on familiarity with a
variety of computer applications,Must have a working knowledge of a
variety of computer software applications in word processing,
spreadsheets, databases, presentation software (MS Word, Excel,
PowerPoint), and OutlookA valid driver’s license is requiredTop Secret
clearance required
Knowledge, Skills, And Abilities
Possess a demonstrated ability to analyze documents to extract
informationGood oral and written communication skillsHave hands-on
familiarity with a variety of computer applications, including word
processing, database, spreadsheet, and telecommunications softwareMust
be a team playerMust be able to work independently and with USMS staff
to interpret data rapidly and accurately for proper execution in a
records management databaseMust have a working knowledge of a variety of
computer software applications in word processing, spreadsheets,
databases, presentation software (MS Word, Excel, Access, PowerPoint),
and OutlookAbility to work independently on tasks be a self-starter and
complete projects with a team as they ariseAttention to detail and the
ability to direct the work of others efficiently and effectivelyAbility
to consistently deliver high-quality work under extreme pressureAbility
to work shiftworkAbility to lift and move boxes up to 25 pounds,
including frequently utilizing hands, arms, and legs for file placement
and removalExperience with scanning software
How You’ll Grow
At Chenega MIOS, our professional development plan focuses on helping
our team members at every level of their career to identify and use
their strengths to do their best work every day. From entry-level
employees to senior leaders, we believe there’s always room to learn.
We offer opportunities to help sharpen skills in addition to hands-on
experience in the global, fast-changing business world. From on-the-job
learning experiences to formal development programs, our professionals
have a variety of opportunities to continue to grow throughout their
careers.
Benefits
At Chenega MIOS, we know that great people make a great organization. We
value our team members and offer them a broad range of benefits.
Learn more about what working at Chenega MIOS can mean for you.
Chenega MIOS’s culture
Our positive and supportive culture encourages our team members to do
their best work every day. We celebrate individuals by recognizing their
uniqueness and offering them the flexibility to make daily choices that
can help them be healthy, centered, confident, and aware. We offer
well-being programs and continuously look for new ways to maintain a
culture where we excel and lead healthy, happy lives.
Corporate citizenship
Chenega MIOS is led by a purpose to make an impact that matters. This
purpose defines who we are and extends to relationships with our
clients, our team members, and our communities. We believe that business
has the power to inspire and transform. We focus on education, giving,
skill-based volunteerism, and leadership to help drive positive social
impact in our communities.
Learn more about Chenega’s impact on the world.
Chenega MIOS News- https://chenegamios.com/news/
Tips from your Talent Acquisition team
We Want Job Seekers Exploring Opportunities At Chenega MIOS To Feel
Prepared And Confident. To Help You With Your Research, We Suggest You
Review The Following Links
Chenega MIOS web site - www.chenegamios.com
Glassdoor -
https://www.glassdoor.com/Overview/Working-at-Chenega-MIOS-EI_IE369514.11,23.htm
LinkedIn - https://www.linkedin.com/company/1472684/
Facebook - https://www.facebook.com/chenegamios/
#DICE
#Chenega Defense & Aerospace Solutions, LLC
- |-
Qualifications
Data Engineering, Data Modeling, and ETL (Extract Transform Load) skillsData Warehousing and Data Analytics skillsExperience with data-related tools and technologiesStrong problem-solving and analytical skillsExcellent written and verbal communication skillsAbility to work independently and remotelyExperience with cloud platforms (e.g., AWS, Azure) is a plusBachelor's degree in Computer Science, Information Systems, or related field
- >-
skills will be difficult. The more aligned skills they have, the
better.Organizational Structure And Impact:Describe the function your
group supports from an LOB perspective:Experienced ML engineer to work
on universal forecasting models. Focus on ML forecasting, Python and
Hadoop. Experience with Python, ARIMA, FB Prophet, Seasonal Naive,
Gluon.Data Science Innovation (DSI) is a very unique application. It is
truly ML-driven at its heart and our forecasting models originally
looked singularly at cash balance forecasting. That has all changed as
we have now incorporated approximately 100 additional financial metrics
from our new DSI Metrics Farm. This allows future model executions to
become a Universal Forecasting Model instead of being limited to just
cash forecasting. It’s a very exciting application, especially since the
models have been integrated within a Marketplace concept UI that allows
Subscriber/Contributor functionality to make information and processing
more personal and with greater extensibility across the enterprise. The
application architecture is represented by OpenShift, Linux, Oracle, SQL
Server, Hadoop, MongoDB, APIs, and a great deal of Python code.Describe
the current initiatives that this resource will be impacting:Working
toward implementation of Machine Learning Services.Team Background and
Preferred Candidate History:Do you only want candidates with a similar
background or would you like to see candidates with a diverse industry
background?Diverse industry background, finance background preferred.
Manager is more focused on the skillset.Describe the dynamic of your
team and where this candidate will fit into the overall environment:This
person will work with a variety of titles including application
architects, web engineers, data engineers, data scientists, application
system managers, system integrators, and Quality Engineers.Will work
with various teams, but primarily working with one core team - approx 15
- onshore and offshore resources.Candidate Technical and skills
profile:Describe the role and the key responsibilities in order of which
they will be doing daily:Machine Learning Engineer that work with Data
Scientists in a SDLC environment into production.Interviews:Describe
interview process (who will be involved, how many interviews, etc.):1
round - 1 hour minimum, panel style
datasets:
- ShushantLLM/ai-job-embedding-finetuning
pipeline_tag: sentence-similarity
library_name: sentence-transformers
metrics:
- cosine_accuracy
model-index:
- name: >-
SentenceTransformer based on
sentence-transformers/paraphrase-multilingual-mpnet-base-v2
results:
- task:
type: triplet
name: Triplet
dataset:
name: ai job validation
type: ai-job-validation
metrics:
- type: cosine_accuracy
value: 0.9801980257034302
name: Cosine Accuracy
- task:
type: triplet
name: Triplet
dataset:
name: ai job test
type: ai-job-test
metrics:
- type: cosine_accuracy
value: 0.9607843160629272
name: Cosine Accuracy
SentenceTransformer based on sentence-transformers/paraphrase-multilingual-mpnet-base-v2
This is a sentence-transformers model finetuned from sentence-transformers/paraphrase-multilingual-mpnet-base-v2 on the ai-job-embedding-finetuning dataset. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
Model Details
Model Description
- Model Type: Sentence Transformer
- Base model: sentence-transformers/paraphrase-multilingual-mpnet-base-v2
- Maximum Sequence Length: 128 tokens
- Output Dimensionality: 768 dimensions
- Similarity Function: Cosine Similarity
- Training Dataset:
Model Sources
- Documentation: Sentence Transformers Documentation
- Repository: Sentence Transformers on GitHub
- Hugging Face: Sentence Transformers on Hugging Face
Full Model Architecture
SentenceTransformer(
(0): Transformer({'max_seq_length': 128, 'do_lower_case': False, 'architecture': 'XLMRobertaModel'})
(1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
)
Usage
Direct Usage (Sentence Transformers)
First install the Sentence Transformers library:
pip install -U sentence-transformers
Then you can load this model and run inference.
from sentence_transformers import SentenceTransformer
# Download from the 🤗 Hub
model = SentenceTransformer("ShushantLLM/paraphrase-multilingual-mpnet-base-v2")
# Run inference
queries = [
"Data organization, document analysis, records management",
]
documents = [
'skills and build your career in a rapidly evolving business climate? Are you looking for a career where professional development is embedded in your employer’s core culture? If so, Chenega Military, Intelligence & Operations Support (MIOS) could be the place for you! Join our team of professionals who support large-scale government operations by leveraging cutting-edge technology and take your career to the next level!\n\nAs one of the newest Chenega companies, Chenega Defense & Aerospace Solutions (CDAS) was developed with the purpose of providing expert Engineering and Technical Support Services to federal customers.\n\nThe Data Analyst will analyze a large variety of documents to ensure proper placement in physical files, perform high-level scanning of master file documents to convert them into an electronic format, and provide meticulous organization and management of case files, including sorting and categorizing documents before scanning.\n\nResponsibilities\n\nWork within the Standard Operating Procedure for the organization of physical files containing documents of various types Establish or maintain physical files, including proper placement of documents as they are createdDisseminate significant amounts of information with attention to detail and accuracyPerform word processing tasksPerform data entry and metadata entry for electronic documentsReconcile inconsistenciesGather information and organize investigative packages, case files, or presentationsObtain additional information from other investigative agencies or databasesVerify information and files against the tracking systemMaintain internal status information on the disposition of designated information and filesDistribute and receive documentsAssist analyst or government official in obtaining or collecting all documents or information to complete case fileProvide administrative information and assistance concerning the case or files to other agencies or organizationsOther duties as assigned\n\n\nQualifications\n\nHigh school diploma or GED equivalent required Must have resided in the United States for at least three out of the last five years or worked for the U.S. in a foreign country as either an employee or contractor in a federal or military capacity for at least three of the last five yearsHaving your own Personally Owned Vehicle (POV) is requiredPossess a demonstrated ability to analyze documents to extract informationGood oral and written communication skillsHave hands-on familiarity with a variety of computer applications,Must have a working knowledge of a variety of computer software applications in word processing, spreadsheets, databases, presentation software (MS Word, Excel, PowerPoint), and OutlookA valid driver’s license is requiredTop Secret clearance required \n\n\nKnowledge, Skills, And Abilities\n\nPossess a demonstrated ability to analyze documents to extract informationGood oral and written communication skillsHave hands-on familiarity with a variety of computer applications, including word processing, database, spreadsheet, and telecommunications softwareMust be a team playerMust be able to work independently and with USMS staff to interpret data rapidly and accurately for proper execution in a records management databaseMust have a working knowledge of a variety of computer software applications in word processing, spreadsheets, databases, presentation software (MS Word, Excel, Access, PowerPoint), and OutlookAbility to work independently on tasks be a self-starter and complete projects with a team as they ariseAttention to detail and the ability to direct the work of others efficiently and effectivelyAbility to consistently deliver high-quality work under extreme pressureAbility to work shiftworkAbility to lift and move boxes up to 25 pounds, including frequently utilizing hands, arms, and legs for file placement and removalExperience with scanning software\n\n\nHow You’ll Grow\n\nAt Chenega MIOS, our professional development plan focuses on helping our team members at every level of their career to identify and use their strengths to do their best work every day. From entry-level employees to senior leaders, we believe there’s always room to learn.\n\nWe offer opportunities to help sharpen skills in addition to hands-on experience in the global, fast-changing business world. From on-the-job learning experiences to formal development programs, our professionals have a variety of opportunities to continue to grow throughout their careers.\n\nBenefits\n\nAt Chenega MIOS, we know that great people make a great organization. We value our team members and offer them a broad range of benefits.\n\nLearn more about what working at Chenega MIOS can mean for you.\n\nChenega MIOS’s culture\n\nOur positive and supportive culture encourages our team members to do their best work every day. We celebrate individuals by recognizing their uniqueness and offering them the flexibility to make daily choices that can help them be healthy, centered, confident, and aware. We offer well-being programs and continuously look for new ways to maintain a culture where we excel and lead healthy, happy lives.\n\nCorporate citizenship\n\nChenega MIOS is led by a purpose to make an impact that matters. This purpose defines who we are and extends to relationships with our clients, our team members, and our communities. We believe that business has the power to inspire and transform. We focus on education, giving, skill-based volunteerism, and leadership to help drive positive social impact in our communities.\n\nLearn more about Chenega’s impact on the world.\n\nChenega MIOS News- https://chenegamios.com/news/\n\nTips from your Talent Acquisition team\n\nWe Want Job Seekers Exploring Opportunities At Chenega MIOS To Feel Prepared And Confident. To Help You With Your Research, We Suggest You Review The Following Links\n\nChenega MIOS web site - www.chenegamios.com\n\nGlassdoor - https://www.glassdoor.com/Overview/Working-at-Chenega-MIOS-EI_IE369514.11,23.htm\n\nLinkedIn - https://www.linkedin.com/company/1472684/\n\nFacebook - https://www.facebook.com/chenegamios/\n\n#DICE\n\n#Chenega Defense & Aerospace Solutions, LLC',
'skills will be difficult. The more aligned skills they have, the better.Organizational Structure And Impact:Describe the function your group supports from an LOB perspective:Experienced ML engineer to work on universal forecasting models. Focus on ML forecasting, Python and Hadoop. Experience with Python, ARIMA, FB Prophet, Seasonal Naive, Gluon.Data Science Innovation (DSI) is a very unique application. It is truly ML-driven at its heart and our forecasting models originally looked singularly at cash balance forecasting. That has all changed as we have now incorporated approximately 100 additional financial metrics from our new DSI Metrics Farm. This allows future model executions to become a Universal Forecasting Model instead of being limited to just cash forecasting. It’s a very exciting application, especially since the models have been integrated within a Marketplace concept UI that allows Subscriber/Contributor functionality to make information and processing more personal and with greater extensibility across the enterprise. The application architecture is represented by OpenShift, Linux, Oracle, SQL Server, Hadoop, MongoDB, APIs, and a great deal of Python code.Describe the current initiatives that this resource will be impacting:Working toward implementation of Machine Learning Services.Team Background and Preferred Candidate History:Do you only want candidates with a similar background or would you like to see candidates with a diverse industry background?Diverse industry background, finance background preferred. Manager is more focused on the skillset.Describe the dynamic of your team and where this candidate will fit into the overall environment:This person will work with a variety of titles including application architects, web engineers, data engineers, data scientists, application system managers, system integrators, and Quality Engineers.Will work with various teams, but primarily working with one core team - approx 15 - onshore and offshore resources.Candidate Technical and skills profile:Describe the role and the key responsibilities in order of which they will be doing daily:Machine Learning Engineer that work with Data Scientists in a SDLC environment into production.Interviews:Describe interview process (who will be involved, how many interviews, etc.):1 round - 1 hour minimum, panel style',
"Qualifications\n Data Engineering, Data Modeling, and ETL (Extract Transform Load) skillsData Warehousing and Data Analytics skillsExperience with data-related tools and technologiesStrong problem-solving and analytical skillsExcellent written and verbal communication skillsAbility to work independently and remotelyExperience with cloud platforms (e.g., AWS, Azure) is a plusBachelor's degree in Computer Science, Information Systems, or related field",
]
query_embeddings = model.encode_query(queries)
document_embeddings = model.encode_document(documents)
print(query_embeddings.shape, document_embeddings.shape)
# [1, 768] [3, 768]
# Get the similarity scores for the embeddings
similarities = model.similarity(query_embeddings, document_embeddings)
print(similarities)
# tensor([[ 0.0065, 0.0405, -0.2204]])
Evaluation
Metrics
Triplet
- Datasets:
ai-job-validationandai-job-test - Evaluated with
TripletEvaluator
| Metric | ai-job-validation | ai-job-test |
|---|---|---|
| cosine_accuracy | 0.9802 | 0.9608 |
Training Details
Training Dataset
ai-job-embedding-finetuning
- Dataset: ai-job-embedding-finetuning at 1de228a
- Size: 810 training samples
- Columns:
query,job_description_pos, andjob_description_neg - Approximate statistics based on the first 810 samples:
query job_description_pos job_description_neg type string string string details - min: 9 tokens
- mean: 17.49 tokens
- max: 43 tokens
- min: 8 tokens
- mean: 121.41 tokens
- max: 128 tokens
- min: 7 tokens
- mean: 122.26 tokens
- max: 128 tokens
- Samples:
query job_description_pos job_description_neg Senior Data Analyst, monitoring systems, dashboard development, statistical analysisexperience where you can also make an impact on your community. While safety is a serious business, we are a supportive team that is optimizing the remote experience to create strong and fulfilling relationships even when we are physically apart. Our group of hard-working employees thrive in a positive and inclusive environment, where a bias towards action is rewarded.
We have raised over $380M in venture capital from investors including Tiger Global, Andreessen Horowitz, Matrix Partners, Meritech Capital Partners, and Initialized Capital. Now surpassing a $3B valuation, Flock is scaling intentionally and seeking the best and brightest to help us meet our goal of reducing crime in the United States by 25% in the next three years.
The Opportunity
As a Senior Data Analyst on the ML team, you will be responsible for extracting insights aggregated from various data sources, developing dashboards to identify trends and patterns that highlight model performance issues, performing analysis...SKILLS and EXPERIENCE:3-5+ years of experience domain knowledge with either support of core Banking application experience, Mortgage Servicing or Loan Originations or personal or auto loans within Finance Industry environmentAble to interact with the VP or C-level Business Executives and higher to gather requirements and collaborate with IT; working effectively and independently as well as be collaborative team-oriented team player.Ideally supported Mortgage servicing systems such as Black Knight’s MSP, Sagent, Finastra’s Fusion Servicing Director, Interlinq Loan Servicing (ILS) or other loan servicing platform OR support of other core banking or originations platformSome experience with the following core technologies: T-SQL; SQL Server 2016 or higher; Visual Studio 2017 or higher; SQL Server Data Tools; Team Foundation ServerWorking knowledge of T-SQL programming and scripting, as well as optimization techniques· 3 years of experience with a strong focus on SQL Relational databases, ...advanced analytics, financial strategy, data visualizationskills and business acumen to drive impactful results that inform strategic decisions.Commitment to iterative development, with a proven ability to engage and update stakeholders bi-weekly or as necessary, ensuring alignment, feedback incorporation, and transparency throughout the project lifecycle.Project ownership and development from inception to completion, encompassing tasks such as gathering detailed requirements, data preparation, model creation, result generation, and data visualization. Develop insights, methods or tools using various analytic methods such as causal-model approaches, predictive modeling, regressions, machine learning, time series analysis, etc.Handle large amounts of data from multiple and disparate sources, employing advanced Python and SQL techniques to ensure efficiency and accuracyUphold the highest standards of data integrity and security, aligning with both internal and external regulatory requirements and compliance protocols
Required Qualifications, C...experience Life at Visa.
Job Description
About the Team:
VISA is the leader in the payment industry and has been for a long time, but we are also quickly transitioning into a technology company that is fostering an environment for applying the newest technology to solve exciting problems in this area. For a payment system to work well, the risk techniques, performance, and scalability are critical. These techniques and systems benefit from big data, data mining, artificial intelligence, machine learning, cloud computing, & many other advance technologies. At VISA, we have all of these. If you want to be on the cutting edge of the payment space, learn fast, and make a big impact, then the Artificial Intelligence Platform team may be an ideal place for you!
Our team needs a Senior Data Engineer with proven knowledge of web application and web service development who will focus on creating new capabilities for the AI Platform while maturing our code base and development processes. You...Clinical Operations data analysis, eTMF, EDC implementation, advanced analytics visualizationrequirements, and objectives for Clinical initiatives Technical SME for system activities for the clinical system(s), enhancements, and integration projects. Coordinates support activities across vendor(s) Systems include but are not limited to eTMF, EDC, CTMS and Analytics Interfaces with external vendors at all levels to manage the relationship and ensure the proper delivery of services Document Data Transfer Agreements for Data Exchange between BioNTech and Data Providers (CRO, Partner Organizations) Document Data Transformation logic and interact with development team to convert business logic into technical details
What you have to offer:
Bachelor’s or higher degree in a scientific discipline (e.g., computer science/information systems, engineering, mathematics, natural sciences, medical, or biomedical science) Extensive experience/knowledge of technologies and trends including Visualizations /Advanced Analytics Outstanding analytical skills and result orientation Ab...Requirements
Typically requires 13+ years of professional experience and 6+ years of diversified leadership, planning, communication, organization, and people motivation skills (or equivalent experience).
Critical Skills
12+ years of experience in a technology role; proven experience in a leadership role, preferably in a large, complex organization.8+ years Data Engineering, Emerging Technology, and Platform Design experience4+ years Leading large data / technical teams – Data Engineering, Solution Architects, and Business Intelligence Engineers, encouraging a culture of innovation, collaboration, and continuous improvement.Hands-on experience building and delivering Enterprise Data SolutionsExtensive market knowledge and experience with cutting edge Data, Analytics, Data Science, ML and AI technologiesExtensive professional experience with ETL, BI & Data AnalyticsExtensive professional experience with Big Data systems, data pipelines and data processingDeep expertise in Data Archit... - Loss:
MultipleNegativesRankingLosswith these parameters:{ "scale": 20.0, "similarity_fct": "cos_sim", "gather_across_devices": false }
Evaluation Dataset
ai-job-embedding-finetuning
- Dataset: ai-job-embedding-finetuning at 1de228a
- Size: 101 evaluation samples
- Columns:
query,job_description_pos, andjob_description_neg - Approximate statistics based on the first 101 samples:
query job_description_pos job_description_neg type string string string details - min: 10 tokens
- mean: 17.83 tokens
- max: 32 tokens
- min: 18 tokens
- mean: 122.03 tokens
- max: 128 tokens
- min: 17 tokens
- mean: 119.95 tokens
- max: 128 tokens
- Samples:
query job_description_pos job_description_neg Azure Data Factory, Databricks, Snowflake architectureSkills: SQL, PySpark, Databricks, Azure Synapse, Azure Data Factory.
Need hands-on coding
Requirements:1. Extensive knowledge of any of the big cloud services - Azure, AWS or GCP with practical implementation (like S3, ADLS, Airflow, ADF, Lamda, BigQuery, EC2, Fabric, Databricks or equivalent)2. Strong Hands-on experience in SQL and Python/PySpark programming knowledge. Should be able to write code during an interview with minimal syntax error.3. Strong foundational and architectural knowledge of any of the data warehouses - Snowflake, Redshift. Synapse etc.4. Should be able to drive and deliver projects with little or no guidance. Take ownership, become a self-learner, and have leadership qualities.experience for yourself, and a better working world for all.
Data Analyst, Technology Consulting - Data & Analytics (Data Governance & Controls) - Financial Services Office (Manager) (Multiple Positions), Ernst & Young U.S. LLP, New York, NY.
Work with clients to transform the way they use and manage data by architecting data strategies, providing end-to-end solutions that focus on improving their data supply chain, reengineering processes, enhancing risk control, and enabling information intelligence by harnessing latest advanced technologies. Solve complex issues and drive growth across financial services. Define data and analytic strategies by performing assessments, recommending remediation strategies/solutions based on aggregated view of identified gaps, and designing/implementing future state data and analytics solutions. Manage and coach diverse teams of professionals with different backgrounds. Manage cross functional teams, to ensure project task and timeline accountability...Big Data Engineer, Spark, Hadoop, AWS GCPSkills • Expertise and hands-on experience on Spark, and Hadoop echo system components – Must Have • Good and hand-on experience* of any of the Cloud (AWS/GCP) – Must Have • Good knowledge of HiveQL & SparkQL – Must Have Good knowledge of Shell script & Java/Scala/python – Good to Have • Good knowledge of SQL – Good to Have • Good knowledge of migration projects on Hadoop – Good to Have • Good Knowledge of one of the Workflow engines like Oozie, Autosys – Good to Have Good knowledge of Agile Development– Good to Have • Passionate about exploring new technologies – Good to Have • Automation approach – Good to Have
Thanks & RegardsShahrukh KhanEmail: [email protected]Requirements: We're looking for a candidate with exceptional proficiency in Google Sheets. This expertise should include manipulating, analyzing, and managing data within Google Sheets. The candidate should be outstanding at extracting business logic from existing reports and implementing it into new ones. Although a basic understanding of SQL for tasks related to data validation and metrics calculations is beneficial, the primary skill we are seeking is proficiency in Google Sheets. This role will involve working across various cross-functional teams, so strong communication skills are essential. The position requires a meticulous eye for detail, a commitment to delivering high-quality results, and above all, exceptional competency in Google Sheets
Google sheet knowledge is preferred.Strong Excel experience without Google will be considered.Data Validation and formulas to extract data are a mustBasic SQL knowledge is required.Strong communications skills are requiredInterview process...Energy policy analysis, regulatory impact modeling, distributed energy resource management.skills, modeling, energy data analysis, and critical thinking are required for a successful candidate. Knowledge of energy systems and distributed solar is required.
Reporting to the Senior Manager of Government Affairs, you will work across different teams to model data to inform policy advocacy. The ability to obtain data from multiple sources, including regulatory or legislative hearings, academic articles, and reports, are fundamental to the role.
A willingness to perform under deadlines and collaborate within an organization is required. Honesty, accountability, and integrity are a must.
Energy Policy & Data Analyst Responsibilities
Support Government Affairs team members with energy policy recommendations based on data modelingEvaluate relevant regulatory or legislative filings and model the impacts to Sunnova’s customers and businessAnalyze program proposals (grid services, incentives, net energy metering, fixed charges) and develop recommendations that align with Sunnova’s ...QualificationsData Engineering, Data Modeling, and ETL (Extract Transform Load) skillsMonitor and support data pipelines and ETL workflowsData Warehousing and Data Analytics skillsExperience with Azure cloud services and toolsStrong problem-solving and analytical skillsProficiency in SQL and other programming languagesExperience with data integration and data migrationExcellent communication and collaboration skillsBachelor's degree in Computer Science, Engineering, or related field
Enterprise Required SkillsPython, Big data, Data warehouse, ETL, Development, azure, Azure Data Factory, Azure Databricks, Azure SQL Server, Snowflake, data pipelines
Top Skills Details1. 3+ years with ETL Development with Azure stack (Azure Data Factory, Azure Databricks, Azure Blob, Azure SQL). 2. 3+ years with Spark, SQL, and Python. This will show up with working with large sets of data in an enterprise environment. 3. Looking for Proactive individuals who have completed projects from start to complet... - Loss:
MultipleNegativesRankingLosswith these parameters:{ "scale": 20.0, "similarity_fct": "cos_sim", "gather_across_devices": false }
Training Hyperparameters
Non-Default Hyperparameters
eval_strategy: stepsper_device_train_batch_size: 16per_device_eval_batch_size: 16learning_rate: 2e-05num_train_epochs: 5warmup_ratio: 0.1batch_sampler: no_duplicates
All Hyperparameters
Click to expand
overwrite_output_dir: Falsedo_predict: Falseeval_strategy: stepsprediction_loss_only: Trueper_device_train_batch_size: 16per_device_eval_batch_size: 16per_gpu_train_batch_size: Noneper_gpu_eval_batch_size: Nonegradient_accumulation_steps: 1eval_accumulation_steps: Nonetorch_empty_cache_steps: Nonelearning_rate: 2e-05weight_decay: 0.0adam_beta1: 0.9adam_beta2: 0.999adam_epsilon: 1e-08max_grad_norm: 1.0num_train_epochs: 5max_steps: -1lr_scheduler_type: linearlr_scheduler_kwargs: {}warmup_ratio: 0.1warmup_steps: 0log_level: passivelog_level_replica: warninglog_on_each_node: Truelogging_nan_inf_filter: Truesave_safetensors: Truesave_on_each_node: Falsesave_only_model: Falserestore_callback_states_from_checkpoint: Falseno_cuda: Falseuse_cpu: Falseuse_mps_device: Falseseed: 42data_seed: Nonejit_mode_eval: Falsebf16: Falsefp16: Falsefp16_opt_level: O1half_precision_backend: autobf16_full_eval: Falsefp16_full_eval: Falsetf32: Nonelocal_rank: 0ddp_backend: Nonetpu_num_cores: Nonetpu_metrics_debug: Falsedebug: []dataloader_drop_last: Falsedataloader_num_workers: 0dataloader_prefetch_factor: Nonepast_index: -1disable_tqdm: Falseremove_unused_columns: Truelabel_names: Noneload_best_model_at_end: Falseignore_data_skip: Falsefsdp: []fsdp_min_num_params: 0fsdp_config: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}fsdp_transformer_layer_cls_to_wrap: Noneaccelerator_config: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}parallelism_config: Nonedeepspeed: Nonelabel_smoothing_factor: 0.0optim: adamw_torch_fusedoptim_args: Noneadafactor: Falsegroup_by_length: Falselength_column_name: lengthproject: huggingfacetrackio_space_id: trackioddp_find_unused_parameters: Noneddp_bucket_cap_mb: Noneddp_broadcast_buffers: Falsedataloader_pin_memory: Truedataloader_persistent_workers: Falseskip_memory_metrics: Trueuse_legacy_prediction_loop: Falsepush_to_hub: Falseresume_from_checkpoint: Nonehub_model_id: Nonehub_strategy: every_savehub_private_repo: Nonehub_always_push: Falsehub_revision: Nonegradient_checkpointing: Falsegradient_checkpointing_kwargs: Noneinclude_inputs_for_metrics: Falseinclude_for_metrics: []eval_do_concat_batches: Truefp16_backend: autopush_to_hub_model_id: Nonepush_to_hub_organization: Nonemp_parameters:auto_find_batch_size: Falsefull_determinism: Falsetorchdynamo: Noneray_scope: lastddp_timeout: 1800torch_compile: Falsetorch_compile_backend: Nonetorch_compile_mode: Noneinclude_tokens_per_second: Falseinclude_num_input_tokens_seen: noneftune_noise_alpha: Noneoptim_target_modules: Nonebatch_eval_metrics: Falseeval_on_start: Falseuse_liger_kernel: Falseliger_kernel_config: Noneeval_use_gather_object: Falseaverage_tokens_across_devices: Trueprompts: Nonebatch_sampler: no_duplicatesmulti_dataset_batch_sampler: proportionalrouter_mapping: {}learning_rate_mapping: {}
Training Logs
| Epoch | Step | Training Loss | Validation Loss | ai-job-validation_cosine_accuracy | ai-job-test_cosine_accuracy |
|---|---|---|---|---|---|
| -1 | -1 | - | - | 0.8416 | - |
| 1.9608 | 100 | 1.2457 | 1.3444 | 0.9802 | - |
| 3.9216 | 200 | 0.3222 | 1.3620 | 0.9802 | - |
| -1 | -1 | - | - | 0.9802 | 0.9608 |
Framework Versions
- Python: 3.12.12
- Sentence Transformers: 5.1.2
- Transformers: 4.57.1
- PyTorch: 2.8.0+cu126
- Accelerate: 1.11.0
- Datasets: 4.0.0
- Tokenizers: 0.22.1
Citation
BibTeX
Sentence Transformers
@inproceedings{reimers-2019-sentence-bert,
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
author = "Reimers, Nils and Gurevych, Iryna",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
month = "11",
year = "2019",
publisher = "Association for Computational Linguistics",
url = "https://arxiv.org/abs/1908.10084",
}
MultipleNegativesRankingLoss
@misc{henderson2017efficient,
title={Efficient Natural Language Response Suggestion for Smart Reply},
author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil},
year={2017},
eprint={1705.00652},
archivePrefix={arXiv},
primaryClass={cs.CL}
}