ShushantLLM's picture
Add new SentenceTransformer model
d493a02 verified
metadata
tags:
  - sentence-transformers
  - sentence-similarity
  - feature-extraction
  - dense
  - generated_from_trainer
  - dataset_size:810
  - loss:MultipleNegativesRankingLoss
base_model: sentence-transformers/paraphrase-multilingual-mpnet-base-v2
widget:
  - source_sentence: CBRN defense, predictive analytics, natural language understanding
    sentences:
      - >-
        experience with speech interfaces Lead and evaluate changing dialog
        evaluation conventions, test tooling developments, and pilot processes
        to support expansion to new data areas Continuously evaluate workflow
        tools and processes and offer solutions to ensure they are efficient,
        high quality, and scalable Provide expert support for a large and
        growing team of data analysts Provide support for ongoing and new data
        collection efforts as a subject matter expert on conventions and use of
        the data Conduct research studies to understand speech and
        customer-Alexa interactions Assist scientists, program and product
        managers, and other stakeholders in defining and validating customer
        experience metrics


        We are open to hiring candidates to work out of one of the following
        locations:


        Boston, MA, USA | Seattle, WA, USA


        Basic Qualifications

         3+ years of data querying languages (e.g. SQL), scripting languages (e.g. Python) or statistical/mathematical software (e.g. R, SAS, Matlab, etc.) experience 2+ years of data scientist experience Bachelor's degree Experience applying theoretical models in an applied environment

        Preferred Qualifications

         Experience in Python, Perl, or another scripting language Experience in a ML or data scientist role with a large technology company Master's degree in a quantitative field such as statistics, mathematics, data science, business analytics, economics, finance, engineering, or computer science

        Amazon is committed to a diverse and inclusive workplace. Amazon is 


        Our compensation reflects the cost of labor across several US geographic
        markets. The base pay for this position ranges from $111,600/year in our
        lowest geographic market up to $212,800/year in our highest geographic
        market. Pay is based on a number of factors including market location
        and may vary depending on job-related knowledge, skills, and experience.
        Amazon is a total compensation company. Dependent on the position
        offered, equity, sign-on payments, and other forms of compensation may
        be provided as part of a total compensation package, in addition to a
        full range of medical, financial, and/or other benefits. For more
        information, please visit
        https://www.aboutamazon.com/workplace/employee-benefits. This position
        will remain posted until filled. Applicants should apply via our
        internal or external career site.



        Company - Amazon.com Services LLC


        Job ID: A2610750
      - >-
        Skills: Your Expertise:

        5+ years in industry experience and a degree (Masters or PhD is a plus)
        in a quantitative field (e.g., Statistics, Econometrics, Computer
        Science, Engineering, Mathematics, Data Science, Operations
        Research).Expert communication and collaboration skills with the ability
        to work effectively with internal teams in a cross-cultural and
        cross-functional environment. Ability to conduct rigorous analysis and
        communicate conclusions to both technical and non-technical
        audiencesExperience partnering with internal teams to drive action and
        providing expertise and direction on analytics, data science,
        experimental design, and measurementExperience in analysis of A|B
        experiments and statistical data analysisExperience designing and
        building metrics, from conception to building prototypes with data
        pipelinesStrong knowledge in at least one programming language (Python
        or R) and in SQLAbility to drive data strategies, with a central source
        of truth to impact business decisionsKnowledge and experience in
        insurance industry - a plusKnowledge and experience in customer
        experience measurement - a plus

        Keywords:Education: Minimum: BS/BA in CS or related field (or
        self-taught/ equivalent work experience) Preferred: MS/MA in CS or
        related field
      - >-
        requirements of the program or company.

         Working across the globe, V2X builds smart solutions designed to integrate physical and digital infrastructure from base to battlefield. We bring 120 years of successful mission support to improve security, streamline logistics, and enhance readiness. Aligned around a shared purpose, our $3.9B company and 16,000 people work alongside our clients, here and abroad, to tackle their most complex challenges with integrity, respect, responsibility, and professionalism. 

        At V2X, we are making a difference by delivering decision support tools
        critical for the protection of our forces when threatened by both
        physical and Chemical, Biological, Radiological, or Nuclear (CBRN)
        threats.


        We are expanding in data science to provide the best information
        possible utilizing the latest techniques in Machine Learning (including
        Deep Learning, Neural network). We are on the forefront of CBRN defense
        and we are looking for talented Data Scientists that have applied
        experience in the fields of artificial intelligence, machine learning
        and/or natural language processing to join our team. Our data scientists
        work closely everyday with project managers, subject matter experts and
        software engineers to work on challenges in machine intelligence, data
        mining, and machine learning, and work together with agility to build
        capabilities that impress our customers. We partner and collaborate with
        universities to being best minds together.


        Data scientists will have opportunities to work on projects with highest
        priority to our business. Vital to success in this role is the ability
        to determine, define and deploy predictive / prescriptive analytic
        solutions to identify and perform root cause analysis on adverse trends,
        by choosing best fit methods, defining algorithms, and validating and
        deploying models to achieve results.


        Responsibilities


        Major Job Activities:

         Partner with our development teams to solve problems and identify trends and opportunities to leverage data from multiple sources.  Collaborate across multiple teams.  Passionate about working with large and complex unstructured and structured data sets.  Strong communication and interpersonal skills. You should be able to work across functions and effectively present, recommend and communicate a position by demonstrating its value and tradeoffs.  Comfortable conducting design, algorithm, and code reviews.  Able to self-direct and succeed with minimal guidance. 

        Material & Equipment Directly Used:


        Computer, Phone, and basic office materials.


        Working Environment:

         Function in an office environment in a stationary position approximately 50 percent of the time or more.  Must be able to operate standard office equipment, such as a computer, copy machine, and printer. 

        Qualifications


        Education / Certifications:

         Bachelor’s degree in a computer, engineering, or quantitative discipline (e.g., statistics, operations research, bioinformatics, economics, computational biology, computer science, mathematics, physics, electrical engineering, industrial engineering).  Master's or Ph.D. in a quantitative discipline preferred. 

        Clearance Requirement: 


        Must have or be able to obtain an active U.S. DoD Secret (or higher)
        level Security Clearance.


        Experience / Skills:

         5+ years of relevant work experience in data analysis or related field. (e.g., statistician, data analyst, data scientist).  Programming experience in one or more of the following: R, MATLAB, C, C++, Java, Python, Scala  Experience in Natural Language Understanding, Computer Vision, Machine Learning, Algorithmic Foundations of Optimization, Data Mining or Machine Intelligence (Artificial Intelligence).  Experience with statistical software (e.g., R, Octave, Julia, MATLAB, pandas) and database languages (e.g., SQL).  Experience with machine learning related open source libraries including, but not limited to: Hadoop, Spark, SciKit-Learn, TensorFlow, etc.  Contribution to research communities and/or efforts, including publishing papers at conferences. 

        We are committed to an inclusive and diverse workplace that values and
        supports the contributions of each individual. This commitment along
        with our common Vision and Values of Integrity, Respect, and
        Responsibility, allows us to leverage differences, encourage innovation
        and expand our success in the global marketplace. V2X is an Equal
        Opportunity /Affirmative Action Employer. All qualified applicants will
        receive consideration for employment without regard to race, color,
        religion, age, sex, national origin, protected veteran status or status
        as an individual with a disability.
  - source_sentence: >-
      Senior Data Analyst Pricing, data product automation, pricing strategy
      analysis
    sentences:
      - >-
        Skills You BringBachelor’s or Master’s Degree in a technology related
        field (e.g. Engineering, Computer Science, etc.) required with 6+ years
        of experienceInformatica Power CenterGood experience with ETL
        technologiesSnaplogicStrong SQLProven data analysis skillsStrong data
        modeling skills doing either Dimensional or Data Vault modelsBasic AWS
        Experience Proven ability to deal with ambiguity and work in fast paced
        environmentExcellent interpersonal and communication skillsExcellent
        collaboration skills to work with multiple teams in the organization
      - >-
        experience, an annualized transactional volume of $140 billion in 2023,
        and approximately 3,200 employees located in 12+ countries, Paysafe
        connects businesses and consumers across 260 payment types in over 40
        currencies around the world. Delivered through an integrated platform,
        Paysafe solutions are geared toward mobile-initiated transactions,
        real-time analytics and the convergence between brick-and-mortar and
        online payments. Further information is available at www.paysafe.com.


        Are you ready to make an impact? Join our team that is inspired by a
        unified vision and propelled by passion.


        Position Summary


        We are looking for a dynamic and flexible, Senior Data Analyst, Pricing
        to support our global Sales and Product organizations with strategic
        planning, analysis, and commercial pricing efforts . As a Senior Data
        Analyst , you will be at the frontier of building our Pricing function
        to drive growth through data and AI-enabled capabilities. This
        opportunity is high visibility for someone hungry to drive the upward
        trajectory of our business and be able to contribute to their efforts in
        the role in our success.


        You will partner with Product Managers to understand their commercial
        needs, then prioritize and work with a cross-functional team to deliver
        pricing strategies and analytics-based solutions to solve and execute
        them. Business outcomes will include sustainable growth in both revenues
        and gross profit.


        This role is based in Jacksonville, Florida and offers a flexible hybrid
        work environment with 3 days in the office and 2 days working remote
        during the work week.


        Responsibilities

         Build data products that power the automation and effectiveness of our pricing function, driving better quality revenues from merchants and consumers.  Partner closely with pricing stakeholders (e.g., Product, Sales, Marketing) to turn raw data into actionable insights. Help ask the right questions and find the answers.  Dive into complex pricing and behavioral data sets, spot trends and make interpretations.  Utilize modelling and data-mining skills to find new insights and opportunities.  Turn findings into plans for new data products or visions for new merchant features.  Partner across merchant Product, Sales, Marketing, Development and Finance to build alignment, engagement and excitement for new products, features and initiatives.  Ensure data quality and integrity by following and enforcing data governance policies, including alignment on data language. 

          Qualifications  

         Bachelor’s degree in a related field of study (Computer Science, Statistics, Mathematics, Engineering, etc.) required.  5+ years of experience of in-depth data analysis role, required; preferably in pricing context with B2B & B2C in a digital environment.  Proven ability to visualize data intuitively, cleanly and clearly in order to make important insights simplified.  Experience across large and complex datasets, including customer behavior, and transactional data.  Advanced in SQL and in Python, preferred.  Experience structuring and analyzing A/B tests, elasticities and interdependencies, preferred.  Excellent communication and presentation skills, with the ability to explain complex data insights to non-technical audiences. 

         Life at Paysafe: 

        One network. One partnership. At Paysafe, this is not only our business
        model; this is our mindset when it comes to our team. Being a part of
        Paysafe means you’ll be one of over 3,200 members of a world-class team
        that drives our business to new heights every day and where we are
        committed to your personal and professional growth.


        Our culture values humility, high trust & autonomy, a desire for
        excellence and meeting commitments, strong team cohesion, a sense of
        urgency, a desire to learn, pragmatically pushing boundaries, and
        accomplishing goals that have a direct business impact.

         

        Paysafe provides equal employment opportunities to all employees, and
        applicants for employment, and prohibits discrimination of any type
        concerning ethnicity, religion, age, sex, national origin, disability
        status, sexual orientation, gender identity or expression, or any other
        protected characteristics. This policy applies to all terms and
        conditions of recruitment and employment. If you need any reasonable
        adjustments, please let us know. We will be happy to help and look
        forward to hearing from you.
      - "Experience : 10 yearsLocation : RemoteDuration: Full TimeJob DetailsData Warehouse, ETL, Advanced SQL,Data Profiling, Source to Target Mapping,Business Requirement Document, FRS, Healthcare.Should be able to navigate the code - developer background\n\uFEFFThanks & Regard's\nMohd FurquanLead Technical RecruiterE-mail: [email protected] No: +1 925 313 8949LinkedIn-ID :linkedin.com/in/mohd-furquan-94237816aVisit us: www.msrcosmos.com"
  - source_sentence: >-
      CPG data analysis, Nielsen IRI expertise, Power Query dashboard
      development
    sentences:
      - >-
        Skills :

        a) Azure Data Factory  Min 3 years of project experiencea. Design of
        pipelinesb. Use of project with On-prem to Cloud Data Migrationc.
        Understanding of ETLd. Change Data Capture from Multiple Sourcese. Job
        Schedulingb) Azure Data Lake  Min 3 years of project experiencea. All
        steps from design to deliverb. Understanding of different Zones and
        design principalc) Data Modeling experience Min 5 Yearsa. Data
        Mart/Warehouseb. Columnar Data design and modelingd) Reporting using
        PowerBI Min 3 yearsa. Analytical Reportingb. Business Domain Modeling
        and data dictionary

        Interested please apply to the job, looking only for W2 candidates.
      - >-
        experienced and highly skilled Sr Data Engineer to join us. This role
        requires a seasoned professional with a deep understanding of automated
        data pipelines, cloud infrastructure, databases, and workflow engines.
        The ideal candidate will have a minimum of 5 years of technical lead
        experience in the medical device field and at least 7 years of
        experience in data engineering. Proficiency in Python and a proven track
        record of leading projects to completion are essential.


        Primary Duties


        Design, develop, and manage robust, secure, scalable, highly available,
        and dynamic solutions to drive business objectives. Lead the
        architecture and implementation of advanced cloud-based data engineering
        solutions, leveraging AWS technologies and best practices. Manage and
        optimize data pipelines, ensuring timely and accurate data availability
        for analytics and machine learning applications. Oversee the
        administration and performance tuning of databases and workflow engines.
        Collaborate with cross-functional teams (e.g., product management, IT,
        software engineering) to define data requirements, integrate systems,
        and implement data governance and security policies. Mentor junior data
        engineers and oversee the team's development efforts, promoting best
        practices in coding, architecture, and data management. Stay abreast of
        emerging technologies and trends in data engineering, cloud services,
        and the medical device industry to drive innovation and competitive
        advantage. 


        Knowledge, Experience & Skills


        Degree in Computer Science, Engineering, Information Systems, or a
        related field. Requiring a minimum of Bachelor’s degree +7yrs of
        experience or a Master’s degree +5yrs of experience. Minimum of 7 years
        of experience in data engineering, with expertise in developing and
        managing automated data pipelines, AWS cloud infrastructure, databases,
        and workflow engines. Certifications in AWS and data engineering
        preferred. Experience with machine learning algorithms and data modeling
        techniques. At least 5 years of experience in the medical device IVD
        industry, with a strong understanding of FDA regulatory standards and
        compliance requirements. Expert proficiency in Python programming and
        software engineering principles. Demonstrated experience with AWS
        services (e.g., EC2, RDS, S3, Lambda, Glue, Redshift, Athena, EMR) and
        data pipeline tools (e.g., Apache Airflow, Luigi, etc). Strong knowledge
        of database management (Postgres and Snowflake), SQL, and NoSQL
        databases. Adept at queries, report writing and presenting findings
        Experienced in developing and maintaining ETL pipelines in a cloud
        environmentExperienced in Unit Testing preferred Strong analytical
        skills with the ability to organize, analyze, and disseminate
        information with attention to detail and accuracy Excellent
        communication and task management skills. Comfort working in a dynamic,
        fast-paced, research-oriented group with several ongoing concurrent
        projectsFull fluency (verbal and written) of the English language is a
        must. 


        The estimated salary range for this role based in California is between
        $148,700 and $178,400 annually. This role is eligible to receive a
        variable annual bonus based on company, team, and individual performance
        per bioMerieux’s bonus program. This range may differ from ranges
        offered for similar positions elsewhere in the country given differences
        in cost of living. Actual compensation within this range is determined
        based on the successful candidate’s experience and will be presented in
        writing at the time of the offer.


        In addition, bioMérieux offers a competitive Total Rewards package that
        may include:


        A choice of medical (including prescription), dental, and vision plans
        providing nationwide coverage and telemedicine optionsCompany-Provided
        Life and Accidental Death InsuranceShort and Long-Term Disability
        InsuranceRetirement Plan including a generous non-discretionary employer
        contribution and employer match. Adoption AssistanceWellness
        ProgramsEmployee Assistance ProgramCommuter BenefitsVarious voluntary
        benefit offeringsDiscount programsParental leaves


        BioFire Diagnostics, LLC. is an Equal Opportunity/Affirmative Action
        Employer. All qualified applicants will receive consideration for
        employment without regard to race, color, religion, sex, sexual
        orientation, gender identity, national origin, age, protected veteran or
        disabled status, or genetic information.


        Please be advised that the receipt of satisfactory responses to
        reference requests and the provision of satisfactory proof of an
        applicant’s identity and legal authorization to work in the United
        States are required of all new hires. Any misrepresentation,
        falsification, or material omission may result in the failure to receive
        an offer, the retraction of an offer, or if already hired, dismissal. If
        you are a qualified individual with a disability, you may request a
        reasonable accommodation in BioFire Diagnostics’ application process by
        contacting us via telephone at (385) 770-1132, by email at [email
        protected], or by dialing 711 for access to Telecommunications Relay
        Services (TRS).
      - >-
        requirements into analytical frameworks.Dashboard Development: Design
        and maintain dashboards using Power Query in Excel, good in analytics in
        generating metrics & measures and ensuring accurate and real-time data
        representation. 

        Required QualificationsProfessional Experience: 3-6 years as a business
        analyst, with mandatory experience in the CPG sector and should have
        worked on brand dataTechnical Proficiency: Advanced skills in Excel and
        Power Query;Communication Skills: Exceptional ability to communicate
        complex data insights to non-technical stakeholders.Location: Position
        based in Springdale. Preferred AttributesProven experience in
        data-driven decision-making processes.Ability to handle multiple
        projects simultaneously, with a focus on deadlines and results.
  - source_sentence: ETL Pipelines, Apache Spark, AirFlow
    sentences:
      - >-
        Qualifications

         - Currently enrolled in a Bachelor’s or Master’s degree in Software Development, Computer Science, Computer Engineering, or a related technical discipline
        - Must obtain work authorization in country of employment at the time of
        hire, and maintain ongoing work authorization during employment.


        Preferred Qualifications: 

        - Fluency in SQL or other programming languages (Python, R etc) for data
        manipulation

        - Ability to thrive in a fast paced work environment 

        - Ability to drive projects to completion with minimal guidance

        - Ability to communicate the results of analyses in a clear and
        effective manner


        TikTok is committed to creating an inclusive space where employees are
        valued for their skills, experiences, and unique perspectives. Our
        platform connects people from across the globe and so does our
        workplace. At TikTok, our mission is to inspire creativity and bring
        joy. To achieve that goal, we are committed to celebrating our diverse
        voices and to creating an environment that reflects the many communities
        we reach. We are passionate about this and hope you are too.


        TikTok is committed to providing reasonable accommodations in our
        recruitment processes for candidates with disabilities, pregnancy,
        sincerely held religious beliefs or other reasons protected by
        applicable laws. If you need assistance or a reasonable accommodation,
        please reach out to us at https://shorturl.at/cdpT2


        By submitting an application for this role, you accept and agree to our
        global applicant privacy policy, which may be accessed here:
        https://careers.tiktok.com/legal/privacy. 


        Job Information:


        【For Pay Transparency】Compensation Description (annually) The base
        salary range for this position in the selected city is $45 -
        $45annually. We cover 100% premium coverage for Full-Time intern medical
        insurance after 90 days from the date of hire. Medical coverage only, no
        dental or vision coverage.Our time off and leave plans are: Paid
        holidays and paid sick leave. The sick leave entitlement is based on the
        time you join.We also provide mental and emotional health benefits
        through our Employee Assistance Program and provide reimbursements for
        your mobile phone expense. The Company reserves the right to modify or
        change these benefits programs at any time, with or without notice.
      - >-
        Experience as a Product Data Analyst at TGG:Achieving business results
        as a client facing consultant for our clients in various types of
        engagements within a variety of industries.Delivering high quality work
        to our clients within our technology service line. Being part of a
        collaborative, values-based firm that has a reputation for great work
        and satisfied clients.Working with senior IT leaders to communicate
        strategic goals to their organization, including leading client and
        internal development teams on best practices.

        What You Will Work On:Analyze large datasets to identify patterns,
        trends, and opportunities for product optimization.Develop and maintain
        dashboards and reports to track key performance metrics.Collaborate with
        product managers, marketers, and engineers to ideate, prioritize, and
        implement data-driven initiatives.Conduct A/B testing and other
        statistical analyses to evaluate the effectiveness of product
        changes.Communicate findings and recommendations to stakeholders through
        clear and concise presentations.Contribute analytical insights to inform
        product vision and deliver value.

        Who Will You Work With:Client stakeholders ranging from individual
        contributors to senior executives.A collaborative team of consultants
        that deliver outstanding client service.TGG partners, principals,
        account leaders, managers, and staff supporting you to excel within
        client projects and to achieve your professional development goals.

        Examples of What You Bring to the Table:You have strong analysis
        capabilities and thrive on working collaboratively to deliver successful
        results for clients. You have experience with these
        technologies:Proficiency in SQL and Python for data extraction,
        manipulation, and analysis.Strong understanding of statistical concepts
        and techniques.Intermediate experience with Tableau, Power BI, Adobe
        Analytics, or similar BI tools.Ability to analyze requirements, design,
        implement, debug, and deploy Cloud Platform services and components.At
        least basic exposure to data science and machine learning
        methods.Familiarity with source control best practices: Define,
        Setup/Configure, Deploy and Maintain source code (e.g. GIT, VisualSafe
        Source).Ability to develop and schedule processes to extract, transform,
        and store data from these systems: SQL databases, Azure cloud services,
        Google cloud service, Snowflake.4-8 years of relevant
        experience.Bachelor’s degree in Computer Science, Statistics, Economics,
        Mathematics, or a related field; or equivalent combination of education,
        training, and experience.Analytical Product Mindset: Ability to approach
        problems analytically and derive actionable insights from complex
        datasets, while remaining focused on providing value to customers
        Strategic Thinking: Demonstrated ability to translate data findings into
        strategic, achievable recommendations to drive business
        outcomes.Communication Skills: Excellent verbal and written
        communication skills.Ability to effectively convey technical concepts
        from technical to non-technical stakeholders and vice-versa.Team Player:
        Proven track record of collaborating effectively with cross-functional
        teams in a fast-paced environment.Adaptability: Have consistently
        demonstrated the ability to bring structure to complex, unstructured
        environments.Familiarity with Agile development methodologies.Ability to
        adapt to changing priorities to thrive in dynamic work environments.

        Salary and Benefits:Nothing is more important to us than the well-being
        of our team. That is why we are proud to offer a full suite of
        competitive health benefits along with additional benefits such as:
        flexible PTO, a professional development stipend and work from home
        stipend, volunteer opportunities, and team social activities.

        Salaries vary and are dependent on considerations such as: experience
        and specific skills/certifications. The base plus target bonus total
        compensation range for this role is $95,000 - $125,000. Additional
        compensation beyond this range is available as a result of leadership
        and business development opportunities. Salary details are discussed
        openly during the hiring process. 

        Work Environment:TGG is headquartered in Portland, Oregon, and has team
        members living in various locations across the United States. Our
        consultants must have the ability to travel and to work remotely or
        onsite. Each engagement has unique conditions, and we work
        collaboratively to meet both our client and team's needs regarding
        onsite and travel requirements. 

        Why The Gunter Group:TGG was created to be different, to be relational,
        to be insightful, and to maximize potential for our consultants, our
        clients, and our community. We listen first so we can learn, analyze,
        and deliver meaningful solutions for our clients. Our compass points
        towards our people and our “Non-Negotiables” always. Our driven
        employees make us who we are  a talented team of leaders with deep and
        diverse professional experience.If you think this role is the right fit,
        please submit your resume and cover letter so we can learn more about
        you. 

        The Gunter Group LLC is
      - >-
        Requirements & Day-to-Day:  Design, develop, and support scalable data
        processing pipelines using Apache Spark and Java/Scala. Lead a talented
        team and make a significant impact on our data engineering capabilities.
        Implement and manage workflow orchestration with AirFlow for efficient
        data processing. Proficiently use SQL for querying and data manipulation
        tasks. Collaborate with cross-functional teams to gather requirements
        and ensure alignment with data engineering solutions.  Essential
        Criteria:  a bachelor’s degree in computer science or another relevant
        discipline, and a minimum of five years of relevant experience in data
        engineering. Solid experience with Apache Spark for large-scale data
        processing. Proficiency in Java or Scala programming languages. Strong
        knowledge of AirFlow for workflow orchestration. Proficient in SQL for
        data querying and manipulation.
  - source_sentence: Data organization, document analysis, records management
    sentences:
      - >-
        skills and build your career in a rapidly evolving business climate? Are
        you looking for a career where professional development is embedded in
        your employer’s core culture? If so, Chenega Military, Intelligence &
        Operations Support (MIOS) could be the place for you! Join our team of
        professionals who support large-scale government operations by
        leveraging cutting-edge technology and take your career to the next
        level!


        As one of the newest Chenega companies, Chenega Defense & Aerospace
        Solutions (CDAS) was developed with the purpose of providing expert
        Engineering and Technical Support Services to federal customers.


        The Data Analyst will analyze a large variety of documents to ensure
        proper placement in physical files, perform high-level scanning of
        master file documents to convert them into an electronic format, and
        provide meticulous organization and management of case files, including
        sorting and categorizing documents before scanning.


        Responsibilities


        Work within the Standard Operating Procedure for the organization of
        physical files containing documents of various types Establish or
        maintain physical files, including proper placement of documents as they
        are createdDisseminate significant amounts of information with attention
        to detail and accuracyPerform word processing tasksPerform data entry
        and metadata entry for electronic documentsReconcile
        inconsistenciesGather information and organize investigative packages,
        case files, or presentationsObtain additional information from other
        investigative agencies or databasesVerify information and files against
        the tracking systemMaintain internal status information on the
        disposition of designated information and filesDistribute and receive
        documentsAssist analyst or government official in obtaining or
        collecting all documents or information to complete case fileProvide
        administrative information and assistance concerning the case or files
        to other agencies or organizationsOther duties as assigned



        Qualifications


        High school diploma or GED equivalent required Must have resided in the
        United States for at least three out of the last five years or worked
        for the U.S. in a foreign country as either an employee or contractor in
        a federal or military capacity for at least three of the last five
        yearsHaving your own Personally Owned Vehicle (POV) is requiredPossess a
        demonstrated ability to analyze documents to extract informationGood
        oral and written communication skillsHave hands-on familiarity with a
        variety of computer applications,Must have a working knowledge of a
        variety of computer software applications in word processing,
        spreadsheets, databases, presentation software (MS Word, Excel,
        PowerPoint), and OutlookA valid driver’s license is requiredTop Secret
        clearance required 



        Knowledge, Skills, And Abilities


        Possess a demonstrated ability to analyze documents to extract
        informationGood oral and written communication skillsHave hands-on
        familiarity with a variety of computer applications, including word
        processing, database, spreadsheet, and telecommunications softwareMust
        be a team playerMust be able to work independently and with USMS staff
        to interpret data rapidly and accurately for proper execution in a
        records management databaseMust have a working knowledge of a variety of
        computer software applications in word processing, spreadsheets,
        databases, presentation software (MS Word, Excel, Access, PowerPoint),
        and OutlookAbility to work independently on tasks be a self-starter and
        complete projects with a team as they ariseAttention to detail and the
        ability to direct the work of others efficiently and effectivelyAbility
        to consistently deliver high-quality work under extreme pressureAbility
        to work shiftworkAbility to lift and move boxes up to 25 pounds,
        including frequently utilizing hands, arms, and legs for file placement
        and removalExperience with scanning software



        How You’ll Grow


        At Chenega MIOS, our professional development plan focuses on helping
        our team members at every level of their career to identify and use
        their strengths to do their best work every day. From entry-level
        employees to senior leaders, we believe there’s always room to learn.


        We offer opportunities to help sharpen skills in addition to hands-on
        experience in the global, fast-changing business world. From on-the-job
        learning experiences to formal development programs, our professionals
        have a variety of opportunities to continue to grow throughout their
        careers.


        Benefits


        At Chenega MIOS, we know that great people make a great organization. We
        value our team members and offer them a broad range of benefits.


        Learn more about what working at Chenega MIOS can mean for you.


        Chenega MIOS’s culture


        Our positive and supportive culture encourages our team members to do
        their best work every day. We celebrate individuals by recognizing their
        uniqueness and offering them the flexibility to make daily choices that
        can help them be healthy, centered, confident, and aware. We offer
        well-being programs and continuously look for new ways to maintain a
        culture where we excel and lead healthy, happy lives.


        Corporate citizenship


        Chenega MIOS is led by a purpose to make an impact that matters. This
        purpose defines who we are and extends to relationships with our
        clients, our team members, and our communities. We believe that business
        has the power to inspire and transform. We focus on education, giving,
        skill-based volunteerism, and leadership to help drive positive social
        impact in our communities.


        Learn more about Chenega’s impact on the world.


        Chenega MIOS News- https://chenegamios.com/news/


        Tips from your Talent Acquisition team


        We Want Job Seekers Exploring Opportunities At Chenega MIOS To Feel
        Prepared And Confident. To Help You With Your Research, We Suggest You
        Review The Following Links


        Chenega MIOS web site - www.chenegamios.com


        Glassdoor -
        https://www.glassdoor.com/Overview/Working-at-Chenega-MIOS-EI_IE369514.11,23.htm


        LinkedIn - https://www.linkedin.com/company/1472684/


        Facebook - https://www.facebook.com/chenegamios/


        #DICE


        #Chenega Defense & Aerospace Solutions, LLC
      - |-
        Qualifications
         Data Engineering, Data Modeling, and ETL (Extract Transform Load) skillsData Warehousing and Data Analytics skillsExperience with data-related tools and technologiesStrong problem-solving and analytical skillsExcellent written and verbal communication skillsAbility to work independently and remotelyExperience with cloud platforms (e.g., AWS, Azure) is a plusBachelor's degree in Computer Science, Information Systems, or related field
      - >-
        skills will be difficult. The more aligned skills they have, the
        better.Organizational Structure And Impact:Describe the function your
        group supports from an LOB perspective:Experienced ML engineer to work
        on universal forecasting models. Focus on ML forecasting, Python and
        Hadoop. Experience with Python, ARIMA, FB Prophet, Seasonal Naive,
        Gluon.Data Science Innovation (DSI) is a very unique application. It is
        truly ML-driven at its heart and our forecasting models originally
        looked singularly at cash balance forecasting. That has all changed as
        we have now incorporated approximately 100 additional financial metrics
        from our new DSI Metrics Farm. This allows future model executions to
        become a Universal Forecasting Model instead of being limited to just
        cash forecasting. It’s a very exciting application, especially since the
        models have been integrated within a Marketplace concept UI that allows
        Subscriber/Contributor functionality to make information and processing
        more personal and with greater extensibility across the enterprise. The
        application architecture is represented by OpenShift, Linux, Oracle, SQL
        Server, Hadoop, MongoDB, APIs, and a great deal of Python code.Describe
        the current initiatives that this resource will be impacting:Working
        toward implementation of Machine Learning Services.Team Background and
        Preferred Candidate History:Do you only want candidates with a similar
        background or would you like to see candidates with a diverse industry
        background?Diverse industry background, finance background preferred.
        Manager is more focused on the skillset.Describe the dynamic of your
        team and where this candidate will fit into the overall environment:This
        person will work with a variety of titles including application
        architects, web engineers, data engineers, data scientists, application
        system managers, system integrators, and Quality Engineers.Will work
        with various teams, but primarily working with one core team - approx 15
        - onshore and offshore resources.Candidate Technical and skills
        profile:Describe the role and the key responsibilities in order of which
        they will be doing daily:Machine Learning Engineer that work with Data
        Scientists in a SDLC environment into production.Interviews:Describe
        interview process (who will be involved, how many interviews, etc.):1
        round - 1 hour minimum, panel style
datasets:
  - ShushantLLM/ai-job-embedding-finetuning
pipeline_tag: sentence-similarity
library_name: sentence-transformers
metrics:
  - cosine_accuracy
model-index:
  - name: >-
      SentenceTransformer based on
      sentence-transformers/paraphrase-multilingual-mpnet-base-v2
    results:
      - task:
          type: triplet
          name: Triplet
        dataset:
          name: ai job validation
          type: ai-job-validation
        metrics:
          - type: cosine_accuracy
            value: 0.9801980257034302
            name: Cosine Accuracy
      - task:
          type: triplet
          name: Triplet
        dataset:
          name: ai job test
          type: ai-job-test
        metrics:
          - type: cosine_accuracy
            value: 0.9607843160629272
            name: Cosine Accuracy

SentenceTransformer based on sentence-transformers/paraphrase-multilingual-mpnet-base-v2

This is a sentence-transformers model finetuned from sentence-transformers/paraphrase-multilingual-mpnet-base-v2 on the ai-job-embedding-finetuning dataset. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.

Model Details

Model Description

Model Sources

Full Model Architecture

SentenceTransformer(
  (0): Transformer({'max_seq_length': 128, 'do_lower_case': False, 'architecture': 'XLMRobertaModel'})
  (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
)

Usage

Direct Usage (Sentence Transformers)

First install the Sentence Transformers library:

pip install -U sentence-transformers

Then you can load this model and run inference.

from sentence_transformers import SentenceTransformer

# Download from the 🤗 Hub
model = SentenceTransformer("ShushantLLM/paraphrase-multilingual-mpnet-base-v2")
# Run inference
queries = [
    "Data organization, document analysis, records management",
]
documents = [
    'skills and build your career in a rapidly evolving business climate? Are you looking for a career where professional development is embedded in your employer’s core culture? If so, Chenega Military, Intelligence & Operations Support (MIOS) could be the place for you! Join our team of professionals who support large-scale government operations by leveraging cutting-edge technology and take your career to the next level!\n\nAs one of the newest Chenega companies, Chenega Defense & Aerospace Solutions (CDAS) was developed with the purpose of providing expert Engineering and Technical Support Services to federal customers.\n\nThe Data Analyst will analyze a large variety of documents to ensure proper placement in physical files, perform high-level scanning of master file documents to convert them into an electronic format, and provide meticulous organization and management of case files, including sorting and categorizing documents before scanning.\n\nResponsibilities\n\nWork within the Standard Operating Procedure for the organization of physical files containing documents of various types Establish or maintain physical files, including proper placement of documents as they are createdDisseminate significant amounts of information with attention to detail and accuracyPerform word processing tasksPerform data entry and metadata entry for electronic documentsReconcile inconsistenciesGather information and organize investigative packages, case files, or presentationsObtain additional information from other investigative agencies or databasesVerify information and files against the tracking systemMaintain internal status information on the disposition of designated information and filesDistribute and receive documentsAssist analyst or government official in obtaining or collecting all documents or information to complete case fileProvide administrative information and assistance concerning the case or files to other agencies or organizationsOther duties as assigned\n\n\nQualifications\n\nHigh school diploma or GED equivalent required Must have resided in the United States for at least three out of the last five years or worked for the U.S. in a foreign country as either an employee or contractor in a federal or military capacity for at least three of the last five yearsHaving your own Personally Owned Vehicle (POV) is requiredPossess a demonstrated ability to analyze documents to extract informationGood oral and written communication skillsHave hands-on familiarity with a variety of computer applications,Must have a working knowledge of a variety of computer software applications in word processing, spreadsheets, databases, presentation software (MS Word, Excel, PowerPoint), and OutlookA valid driver’s license is requiredTop Secret clearance required \n\n\nKnowledge, Skills, And Abilities\n\nPossess a demonstrated ability to analyze documents to extract informationGood oral and written communication skillsHave hands-on familiarity with a variety of computer applications, including word processing, database, spreadsheet, and telecommunications softwareMust be a team playerMust be able to work independently and with USMS staff to interpret data rapidly and accurately for proper execution in a records management databaseMust have a working knowledge of a variety of computer software applications in word processing, spreadsheets, databases, presentation software (MS Word, Excel, Access, PowerPoint), and OutlookAbility to work independently on tasks be a self-starter and complete projects with a team as they ariseAttention to detail and the ability to direct the work of others efficiently and effectivelyAbility to consistently deliver high-quality work under extreme pressureAbility to work shiftworkAbility to lift and move boxes up to 25 pounds, including frequently utilizing hands, arms, and legs for file placement and removalExperience with scanning software\n\n\nHow You’ll Grow\n\nAt Chenega MIOS, our professional development plan focuses on helping our team members at every level of their career to identify and use their strengths to do their best work every day. From entry-level employees to senior leaders, we believe there’s always room to learn.\n\nWe offer opportunities to help sharpen skills in addition to hands-on experience in the global, fast-changing business world. From on-the-job learning experiences to formal development programs, our professionals have a variety of opportunities to continue to grow throughout their careers.\n\nBenefits\n\nAt Chenega MIOS, we know that great people make a great organization. We value our team members and offer them a broad range of benefits.\n\nLearn more about what working at Chenega MIOS can mean for you.\n\nChenega MIOS’s culture\n\nOur positive and supportive culture encourages our team members to do their best work every day. We celebrate individuals by recognizing their uniqueness and offering them the flexibility to make daily choices that can help them be healthy, centered, confident, and aware. We offer well-being programs and continuously look for new ways to maintain a culture where we excel and lead healthy, happy lives.\n\nCorporate citizenship\n\nChenega MIOS is led by a purpose to make an impact that matters. This purpose defines who we are and extends to relationships with our clients, our team members, and our communities. We believe that business has the power to inspire and transform. We focus on education, giving, skill-based volunteerism, and leadership to help drive positive social impact in our communities.\n\nLearn more about Chenega’s impact on the world.\n\nChenega MIOS News- https://chenegamios.com/news/\n\nTips from your Talent Acquisition team\n\nWe Want Job Seekers Exploring Opportunities At Chenega MIOS To Feel Prepared And Confident. To Help You With Your Research, We Suggest You Review The Following Links\n\nChenega MIOS web site - www.chenegamios.com\n\nGlassdoor - https://www.glassdoor.com/Overview/Working-at-Chenega-MIOS-EI_IE369514.11,23.htm\n\nLinkedIn - https://www.linkedin.com/company/1472684/\n\nFacebook - https://www.facebook.com/chenegamios/\n\n#DICE\n\n#Chenega Defense & Aerospace Solutions, LLC',
    'skills will be difficult. The more aligned skills they have, the better.Organizational Structure And Impact:Describe the function your group supports from an LOB perspective:Experienced ML engineer to work on universal forecasting models. Focus on ML forecasting, Python and Hadoop. Experience with Python, ARIMA, FB Prophet, Seasonal Naive, Gluon.Data Science Innovation (DSI) is a very unique application. It is truly ML-driven at its heart and our forecasting models originally looked singularly at cash balance forecasting. That has all changed as we have now incorporated approximately 100 additional financial metrics from our new DSI Metrics Farm. This allows future model executions to become a Universal Forecasting Model instead of being limited to just cash forecasting. It’s a very exciting application, especially since the models have been integrated within a Marketplace concept UI that allows Subscriber/Contributor functionality to make information and processing more personal and with greater extensibility across the enterprise. The application architecture is represented by OpenShift, Linux, Oracle, SQL Server, Hadoop, MongoDB, APIs, and a great deal of Python code.Describe the current initiatives that this resource will be impacting:Working toward implementation of Machine Learning Services.Team Background and Preferred Candidate History:Do you only want candidates with a similar background or would you like to see candidates with a diverse industry background?Diverse industry background, finance background preferred. Manager is more focused on the skillset.Describe the dynamic of your team and where this candidate will fit into the overall environment:This person will work with a variety of titles including application architects, web engineers, data engineers, data scientists, application system managers, system integrators, and Quality Engineers.Will work with various teams, but primarily working with one core team - approx 15 - onshore and offshore resources.Candidate Technical and skills profile:Describe the role and the key responsibilities in order of which they will be doing daily:Machine Learning Engineer that work with Data Scientists in a SDLC environment into production.Interviews:Describe interview process (who will be involved, how many interviews, etc.):1 round - 1 hour minimum, panel style',
    "Qualifications\n Data Engineering, Data Modeling, and ETL (Extract Transform Load) skillsData Warehousing and Data Analytics skillsExperience with data-related tools and technologiesStrong problem-solving and analytical skillsExcellent written and verbal communication skillsAbility to work independently and remotelyExperience with cloud platforms (e.g., AWS, Azure) is a plusBachelor's degree in Computer Science, Information Systems, or related field",
]
query_embeddings = model.encode_query(queries)
document_embeddings = model.encode_document(documents)
print(query_embeddings.shape, document_embeddings.shape)
# [1, 768] [3, 768]

# Get the similarity scores for the embeddings
similarities = model.similarity(query_embeddings, document_embeddings)
print(similarities)
# tensor([[ 0.0065,  0.0405, -0.2204]])

Evaluation

Metrics

Triplet

Metric ai-job-validation ai-job-test
cosine_accuracy 0.9802 0.9608

Training Details

Training Dataset

ai-job-embedding-finetuning

  • Dataset: ai-job-embedding-finetuning at 1de228a
  • Size: 810 training samples
  • Columns: query, job_description_pos, and job_description_neg
  • Approximate statistics based on the first 810 samples:
    query job_description_pos job_description_neg
    type string string string
    details
    • min: 9 tokens
    • mean: 17.49 tokens
    • max: 43 tokens
    • min: 8 tokens
    • mean: 121.41 tokens
    • max: 128 tokens
    • min: 7 tokens
    • mean: 122.26 tokens
    • max: 128 tokens
  • Samples:
    query job_description_pos job_description_neg
    Senior Data Analyst, monitoring systems, dashboard development, statistical analysis experience where you can also make an impact on your community. While safety is a serious business, we are a supportive team that is optimizing the remote experience to create strong and fulfilling relationships even when we are physically apart. Our group of hard-working employees thrive in a positive and inclusive environment, where a bias towards action is rewarded.

    We have raised over $380M in venture capital from investors including Tiger Global, Andreessen Horowitz, Matrix Partners, Meritech Capital Partners, and Initialized Capital. Now surpassing a $3B valuation, Flock is scaling intentionally and seeking the best and brightest to help us meet our goal of reducing crime in the United States by 25% in the next three years.

    The Opportunity

    As a Senior Data Analyst on the ML team, you will be responsible for extracting insights aggregated from various data sources, developing dashboards to identify trends and patterns that highlight model performance issues, performing analysis...
    SKILLS and EXPERIENCE:3-5+ years of experience domain knowledge with either support of core Banking application experience, Mortgage Servicing or Loan Originations or personal or auto loans within Finance Industry environmentAble to interact with the VP or C-level Business Executives and higher to gather requirements and collaborate with IT; working effectively and independently as well as be collaborative team-oriented team player.Ideally supported Mortgage servicing systems such as Black Knight’s MSP, Sagent, Finastra’s Fusion Servicing Director, Interlinq Loan Servicing (ILS) or other loan servicing platform OR support of other core banking or originations platformSome experience with the following core technologies: T-SQL; SQL Server 2016 or higher; Visual Studio 2017 or higher; SQL Server Data Tools; Team Foundation ServerWorking knowledge of T-SQL programming and scripting, as well as optimization techniques· 3 years of experience with a strong focus on SQL Relational databases, ...
    advanced analytics, financial strategy, data visualization skills and business acumen to drive impactful results that inform strategic decisions.Commitment to iterative development, with a proven ability to engage and update stakeholders bi-weekly or as necessary, ensuring alignment, feedback incorporation, and transparency throughout the project lifecycle.Project ownership and development from inception to completion, encompassing tasks such as gathering detailed requirements, data preparation, model creation, result generation, and data visualization. Develop insights, methods or tools using various analytic methods such as causal-model approaches, predictive modeling, regressions, machine learning, time series analysis, etc.Handle large amounts of data from multiple and disparate sources, employing advanced Python and SQL techniques to ensure efficiency and accuracyUphold the highest standards of data integrity and security, aligning with both internal and external regulatory requirements and compliance protocols

    Required Qualifications, C...
    experience Life at Visa.

    Job Description

    About the Team:

    VISA is the leader in the payment industry and has been for a long time, but we are also quickly transitioning into a technology company that is fostering an environment for applying the newest technology to solve exciting problems in this area. For a payment system to work well, the risk techniques, performance, and scalability are critical. These techniques and systems benefit from big data, data mining, artificial intelligence, machine learning, cloud computing, & many other advance technologies. At VISA, we have all of these. If you want to be on the cutting edge of the payment space, learn fast, and make a big impact, then the Artificial Intelligence Platform team may be an ideal place for you!

    Our team needs a Senior Data Engineer with proven knowledge of web application and web service development who will focus on creating new capabilities for the AI Platform while maturing our code base and development processes. You...
    Clinical Operations data analysis, eTMF, EDC implementation, advanced analytics visualization requirements, and objectives for Clinical initiatives Technical SME for system activities for the clinical system(s), enhancements, and integration projects. Coordinates support activities across vendor(s) Systems include but are not limited to eTMF, EDC, CTMS and Analytics Interfaces with external vendors at all levels to manage the relationship and ensure the proper delivery of services Document Data Transfer Agreements for Data Exchange between BioNTech and Data Providers (CRO, Partner Organizations) Document Data Transformation logic and interact with development team to convert business logic into technical details

    What you have to offer:

    Bachelor’s or higher degree in a scientific discipline (e.g., computer science/information systems, engineering, mathematics, natural sciences, medical, or biomedical science) Extensive experience/knowledge of technologies and trends including Visualizations /Advanced Analytics Outstanding analytical skills and result orientation Ab...
    Requirements

    Typically requires 13+ years of professional experience and 6+ years of diversified leadership, planning, communication, organization, and people motivation skills (or equivalent experience).

    Critical Skills

    12+ years of experience in a technology role; proven experience in a leadership role, preferably in a large, complex organization.8+ years Data Engineering, Emerging Technology, and Platform Design experience4+ years Leading large data / technical teams – Data Engineering, Solution Architects, and Business Intelligence Engineers, encouraging a culture of innovation, collaboration, and continuous improvement.Hands-on experience building and delivering Enterprise Data SolutionsExtensive market knowledge and experience with cutting edge Data, Analytics, Data Science, ML and AI technologiesExtensive professional experience with ETL, BI & Data AnalyticsExtensive professional experience with Big Data systems, data pipelines and data processingDeep expertise in Data Archit...
  • Loss: MultipleNegativesRankingLoss with these parameters:
    {
        "scale": 20.0,
        "similarity_fct": "cos_sim",
        "gather_across_devices": false
    }
    

Evaluation Dataset

ai-job-embedding-finetuning

  • Dataset: ai-job-embedding-finetuning at 1de228a
  • Size: 101 evaluation samples
  • Columns: query, job_description_pos, and job_description_neg
  • Approximate statistics based on the first 101 samples:
    query job_description_pos job_description_neg
    type string string string
    details
    • min: 10 tokens
    • mean: 17.83 tokens
    • max: 32 tokens
    • min: 18 tokens
    • mean: 122.03 tokens
    • max: 128 tokens
    • min: 17 tokens
    • mean: 119.95 tokens
    • max: 128 tokens
  • Samples:
    query job_description_pos job_description_neg
    Azure Data Factory, Databricks, Snowflake architecture Skills: SQL, PySpark, Databricks, Azure Synapse, Azure Data Factory.
    Need hands-on coding
    Requirements:1. Extensive knowledge of any of the big cloud services - Azure, AWS or GCP with practical implementation (like S3, ADLS, Airflow, ADF, Lamda, BigQuery, EC2, Fabric, Databricks or equivalent)2. Strong Hands-on experience in SQL and Python/PySpark programming knowledge. Should be able to write code during an interview with minimal syntax error.3. Strong foundational and architectural knowledge of any of the data warehouses - Snowflake, Redshift. Synapse etc.4. Should be able to drive and deliver projects with little or no guidance. Take ownership, become a self-learner, and have leadership qualities.
    experience for yourself, and a better working world for all.

    Data Analyst, Technology Consulting - Data & Analytics (Data Governance & Controls) - Financial Services Office (Manager) (Multiple Positions), Ernst & Young U.S. LLP, New York, NY.

    Work with clients to transform the way they use and manage data by architecting data strategies, providing end-to-end solutions that focus on improving their data supply chain, reengineering processes, enhancing risk control, and enabling information intelligence by harnessing latest advanced technologies. Solve complex issues and drive growth across financial services. Define data and analytic strategies by performing assessments, recommending remediation strategies/solutions based on aggregated view of identified gaps, and designing/implementing future state data and analytics solutions. Manage and coach diverse teams of professionals with different backgrounds. Manage cross functional teams, to ensure project task and timeline accountability...
    Big Data Engineer, Spark, Hadoop, AWS GCP Skills • Expertise and hands-on experience on Spark, and Hadoop echo system components – Must Have • Good and hand-on experience* of any of the Cloud (AWS/GCP) – Must Have • Good knowledge of HiveQL & SparkQL – Must Have Good knowledge of Shell script & Java/Scala/python – Good to Have • Good knowledge of SQL – Good to Have • Good knowledge of migration projects on Hadoop – Good to Have • Good Knowledge of one of the Workflow engines like Oozie, Autosys – Good to Have Good knowledge of Agile Development– Good to Have • Passionate about exploring new technologies – Good to Have • Automation approach – Good to Have
    Thanks & RegardsShahrukh KhanEmail: [email protected]
    Requirements: We're looking for a candidate with exceptional proficiency in Google Sheets. This expertise should include manipulating, analyzing, and managing data within Google Sheets. The candidate should be outstanding at extracting business logic from existing reports and implementing it into new ones. Although a basic understanding of SQL for tasks related to data validation and metrics calculations is beneficial, the primary skill we are seeking is proficiency in Google Sheets. This role will involve working across various cross-functional teams, so strong communication skills are essential. The position requires a meticulous eye for detail, a commitment to delivering high-quality results, and above all, exceptional competency in Google Sheets

    Google sheet knowledge is preferred.Strong Excel experience without Google will be considered.Data Validation and formulas to extract data are a mustBasic SQL knowledge is required.Strong communications skills are requiredInterview process...
    Energy policy analysis, regulatory impact modeling, distributed energy resource management. skills, modeling, energy data analysis, and critical thinking are required for a successful candidate. Knowledge of energy systems and distributed solar is required.

    Reporting to the Senior Manager of Government Affairs, you will work across different teams to model data to inform policy advocacy. The ability to obtain data from multiple sources, including regulatory or legislative hearings, academic articles, and reports, are fundamental to the role.

    A willingness to perform under deadlines and collaborate within an organization is required. Honesty, accountability, and integrity are a must.

    Energy Policy & Data Analyst Responsibilities

    Support Government Affairs team members with energy policy recommendations based on data modelingEvaluate relevant regulatory or legislative filings and model the impacts to Sunnova’s customers and businessAnalyze program proposals (grid services, incentives, net energy metering, fixed charges) and develop recommendations that align with Sunnova’s ...
    QualificationsData Engineering, Data Modeling, and ETL (Extract Transform Load) skillsMonitor and support data pipelines and ETL workflowsData Warehousing and Data Analytics skillsExperience with Azure cloud services and toolsStrong problem-solving and analytical skillsProficiency in SQL and other programming languagesExperience with data integration and data migrationExcellent communication and collaboration skillsBachelor's degree in Computer Science, Engineering, or related field
    Enterprise Required SkillsPython, Big data, Data warehouse, ETL, Development, azure, Azure Data Factory, Azure Databricks, Azure SQL Server, Snowflake, data pipelines
    Top Skills Details1. 3+ years with ETL Development with Azure stack (Azure Data Factory, Azure Databricks, Azure Blob, Azure SQL). 2. 3+ years with Spark, SQL, and Python. This will show up with working with large sets of data in an enterprise environment. 3. Looking for Proactive individuals who have completed projects from start to complet...
  • Loss: MultipleNegativesRankingLoss with these parameters:
    {
        "scale": 20.0,
        "similarity_fct": "cos_sim",
        "gather_across_devices": false
    }
    

Training Hyperparameters

Non-Default Hyperparameters

  • eval_strategy: steps
  • per_device_train_batch_size: 16
  • per_device_eval_batch_size: 16
  • learning_rate: 2e-05
  • num_train_epochs: 5
  • warmup_ratio: 0.1
  • batch_sampler: no_duplicates

All Hyperparameters

Click to expand
  • overwrite_output_dir: False
  • do_predict: False
  • eval_strategy: steps
  • prediction_loss_only: True
  • per_device_train_batch_size: 16
  • per_device_eval_batch_size: 16
  • per_gpu_train_batch_size: None
  • per_gpu_eval_batch_size: None
  • gradient_accumulation_steps: 1
  • eval_accumulation_steps: None
  • torch_empty_cache_steps: None
  • learning_rate: 2e-05
  • weight_decay: 0.0
  • adam_beta1: 0.9
  • adam_beta2: 0.999
  • adam_epsilon: 1e-08
  • max_grad_norm: 1.0
  • num_train_epochs: 5
  • max_steps: -1
  • lr_scheduler_type: linear
  • lr_scheduler_kwargs: {}
  • warmup_ratio: 0.1
  • warmup_steps: 0
  • log_level: passive
  • log_level_replica: warning
  • log_on_each_node: True
  • logging_nan_inf_filter: True
  • save_safetensors: True
  • save_on_each_node: False
  • save_only_model: False
  • restore_callback_states_from_checkpoint: False
  • no_cuda: False
  • use_cpu: False
  • use_mps_device: False
  • seed: 42
  • data_seed: None
  • jit_mode_eval: False
  • bf16: False
  • fp16: False
  • fp16_opt_level: O1
  • half_precision_backend: auto
  • bf16_full_eval: False
  • fp16_full_eval: False
  • tf32: None
  • local_rank: 0
  • ddp_backend: None
  • tpu_num_cores: None
  • tpu_metrics_debug: False
  • debug: []
  • dataloader_drop_last: False
  • dataloader_num_workers: 0
  • dataloader_prefetch_factor: None
  • past_index: -1
  • disable_tqdm: False
  • remove_unused_columns: True
  • label_names: None
  • load_best_model_at_end: False
  • ignore_data_skip: False
  • fsdp: []
  • fsdp_min_num_params: 0
  • fsdp_config: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
  • fsdp_transformer_layer_cls_to_wrap: None
  • accelerator_config: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
  • parallelism_config: None
  • deepspeed: None
  • label_smoothing_factor: 0.0
  • optim: adamw_torch_fused
  • optim_args: None
  • adafactor: False
  • group_by_length: False
  • length_column_name: length
  • project: huggingface
  • trackio_space_id: trackio
  • ddp_find_unused_parameters: None
  • ddp_bucket_cap_mb: None
  • ddp_broadcast_buffers: False
  • dataloader_pin_memory: True
  • dataloader_persistent_workers: False
  • skip_memory_metrics: True
  • use_legacy_prediction_loop: False
  • push_to_hub: False
  • resume_from_checkpoint: None
  • hub_model_id: None
  • hub_strategy: every_save
  • hub_private_repo: None
  • hub_always_push: False
  • hub_revision: None
  • gradient_checkpointing: False
  • gradient_checkpointing_kwargs: None
  • include_inputs_for_metrics: False
  • include_for_metrics: []
  • eval_do_concat_batches: True
  • fp16_backend: auto
  • push_to_hub_model_id: None
  • push_to_hub_organization: None
  • mp_parameters:
  • auto_find_batch_size: False
  • full_determinism: False
  • torchdynamo: None
  • ray_scope: last
  • ddp_timeout: 1800
  • torch_compile: False
  • torch_compile_backend: None
  • torch_compile_mode: None
  • include_tokens_per_second: False
  • include_num_input_tokens_seen: no
  • neftune_noise_alpha: None
  • optim_target_modules: None
  • batch_eval_metrics: False
  • eval_on_start: False
  • use_liger_kernel: False
  • liger_kernel_config: None
  • eval_use_gather_object: False
  • average_tokens_across_devices: True
  • prompts: None
  • batch_sampler: no_duplicates
  • multi_dataset_batch_sampler: proportional
  • router_mapping: {}
  • learning_rate_mapping: {}

Training Logs

Epoch Step Training Loss Validation Loss ai-job-validation_cosine_accuracy ai-job-test_cosine_accuracy
-1 -1 - - 0.8416 -
1.9608 100 1.2457 1.3444 0.9802 -
3.9216 200 0.3222 1.3620 0.9802 -
-1 -1 - - 0.9802 0.9608

Framework Versions

  • Python: 3.12.12
  • Sentence Transformers: 5.1.2
  • Transformers: 4.57.1
  • PyTorch: 2.8.0+cu126
  • Accelerate: 1.11.0
  • Datasets: 4.0.0
  • Tokenizers: 0.22.1

Citation

BibTeX

Sentence Transformers

@inproceedings{reimers-2019-sentence-bert,
    title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
    author = "Reimers, Nils and Gurevych, Iryna",
    booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
    month = "11",
    year = "2019",
    publisher = "Association for Computational Linguistics",
    url = "https://arxiv.org/abs/1908.10084",
}

MultipleNegativesRankingLoss

@misc{henderson2017efficient,
    title={Efficient Natural Language Response Suggestion for Smart Reply},
    author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil},
    year={2017},
    eprint={1705.00652},
    archivePrefix={arXiv},
    primaryClass={cs.CL}
}