23Apr

Chief Information Security Officer at Merchants Bonding


Job Title:  Chief Information Security Officer (CISO)  

Department:  Information Technology

Position reports to:  Chief Information Officer

Positions that report to this position:  

 

DOL Classification:    Exempt

 

Job Purpose:

The CISO is the primary point-of-contact for information security and will brief CIO and other executives regularly on current cybersecurity threats and challenges, and the status of the information security management systems aiming to improve information security continuously and measurably across the organization. The CISO implements, oversees, and audits cybersecurity identification, protection, detection, response, and recovery processes and systems, with emphasis on IT Security and Compliance.

Job Responsibilities:

Security Strategy Development & Implementation:

  • Working knowledge of system development life cycle in software development, source code development, hardware integration, software virtualization, system administrator, network infrastructure implementation, and database administration. Experience with internal and cloud based hosting and technology delivery models.
  • Develops, implements, and monitors a strategic, comprehensive enterprise information security and IT risk management program to ensure the integrity, confidentiality, and availability of information assets.
  • Applies a broad range of advanced expertise of technology and security principles, best practices, policies and procedures to guide other technology staff in the completion of difficult and complex assignments crossing multiple functional areas.
  • Directs and oversees business continuity planning, disaster recovery; exercising each plan, updating, and modifying with each business partner to ensure positive outcomes.
  • Develops and implements information security policies, standards, guidelines, and procedures.
  • Works collaboratively with technology management peers to ensure security is appropriately included in all development and maintenance activities, including automated security testing where appropriate
  • Regularly update the executive team and the board of directors on the status of information security, risks, and projects.

Cybersecurity management:

  • Working with the cyber security and development team members, evaluates new cybersecurity threats and IT systems for vulnerabilities and malicious activities and to identify potential controls to mitigate them
  • Helps determine the vulnerabilities, risks, impacts and remediation approaches on web development and other software tool and systems.
  • Stays current with the cybersecurity threats and adapts current processes and technologies to mitigate risks to the organization
  • Investigates any breaches and other security incidents that occur

Risk Management:

  • Performs regular security assessments, audits, and relevant risk assessments, communicating results to executive management and board members.  These activities will be done internally as well as through third parties.
  • Leads the incident response team in effectively mitigating security incidents, including data breaches and cyber-attacks, and collaborate with the legal department for any necessary follow-up.
  • Ensures that the company is in compliance with all relevant legal and regulatory requirements, including data protection laws and cybersecurity standards and any industry-specific guidelines.
  • Develops and manages a security awareness and training program for all associates.
  • Ensures that IT security is integrated into the third-party vendor risk management process.

Qualifications: 

Education and Experience: Bachelor’s degree in MIS, CS or equivalent work experience plus 10+ years of progressively more responsibility in cyber security leadership roles.  Strong technical background with knowledge of web development, various software tools, hosting models, hardware platforms and third party technology vendors.

Skills and Knowledge: Project management, attention to detail, problem solving, customer service, critical thinking, above average verbal and written communication skills, ability to explain complex concepts in simple and understandable terms, and collaboration, ability to work independently or work effectively through others, public speaking, and rapport building.

Essential Functions: 

This person must be able to understand and analyze complex business processes.  They must be able to communicate verbally with tact and diplomacy and think on their feet.  They must be able to learn new technology and grasp business concepts quickly in order to synthesize solutions.  This person must be able to make decisions under pressure, problem-solve and troubleshoot complex technology systems

Physical Demands:

The physical demands described here are representative of those that must be met by an associate to successfully perform the essential functions of the job.  Reasonable accommodations may be made to enable individual with disabilities to perform the essential functions. 

The associate may travel up to 10% of the time by air or car to other office or business locations. 

Working Conditions:

The work environment characteristics described here are representative of those an associate encounters while performing the essential functions of this job.  Reasonable accommodations may be made to enable individuals with disabilities to perform the essential functions. 

While performing the duties of this job, the associate is exposed to normal office working conditions that are climate controlled.  The associate would not be exposed to extreme heat or cold or other workplace hazards.  The noise level in the work environment is usually low.

This job description is not intended to be an exhaustive list of skills, efforts, duties, responsibilities or working conditions associated with this position.  The CISO assumes other responsibilities as assigned by the Chief Information Officer.



Source link

23Apr

Japanese Heron-Bench: A Novel AI Benchmark for Evaluating Japanese Capabilities of Vision Language Models VLMs


The rapid progression of Large Language Models (LLMs) is a pivotal milestone in the evolution of artificial intelligence. In recent years, we have witnessed a surge in the development and public accessibility of well-trained LLMs in English and other languages, including Japanese. This expansion underscores a global effort to democratize AI capabilities across linguistic and cultural boundaries.

Building upon the advancements in LLMs, novel approaches have emerged for constructing Vision Language Models (VLMs), which integrate image encoders into language models. These VLMs hold promise in their capacity to understand and generate textual descriptions of visual content. Various evaluation metrics have been proposed to gauge their effectiveness, encompassing tasks such as image captioning, similarity scoring between images and text, and visual question answering (VQA). However, it’s notable that most high-performing VLMs are trained and evaluated predominantly on English-centric datasets.

The need for robust evaluation methodologies becomes increasingly urgent as the demand for non-English models burgeons, particularly in languages like Japanese. Recognizing this imperative, a new evaluation benchmark called the Japanese Heron-Bench has been introduced. This benchmark comprises a meticulously curated dataset of images and contextually relevant questions tailored to the Japanese language and culture. Through this benchmark, the efficacy of VLMs in comprehending visual scenes and responding to queries within the Japanese context can be thoroughly scrutinized.

In tandem with establishing the Japanese Heron-Bench, efforts have been directed toward developing Japanese VLMs trained on Japanese image-text pairs using existing Japanese LLMs. This serves as a foundational step in bridging the gap between LLMs and VLMs in the Japanese linguistic landscape. Such models’ availability facilitates research and fosters innovation in diverse applications ranging from language understanding to visual comprehension.

Despite the strides made in evaluation methodologies, inherent limitations persist. For instance, the accuracy of assessments may be compromised by the performance disparities between languages in LLMs. This is particularly salient in the case of Japanese, where the language proficiency of models may differ from that of English. Additionally, concerns regarding safety aspects such as misinformation, bias, or toxicity in generated content warrant further exploration in evaluation metrics.

In conclusion, while introducing the Japanese Heron-Bench and Japanese VLMs represents significant strides toward comprehensive evaluation and utilization of VLMs in non-English contexts, challenges remain to be addressed. In the future, researchers will research evaluation metrics, and safety considerations will be pivotal in ensuring VLMs’ efficacy, reliability, and ethical deployment across diverse linguistic and cultural landscapes.


Check out the Paper and GithubAll credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter. Join our Telegram Channel, Discord Channel, and LinkedIn Group.

If you like our work, you will love our newsletter..

Don’t Forget to join our 40k+ ML SubReddit


Arshad is an intern at MarktechPost. He is currently pursuing his Int. MSc Physics from the Indian Institute of Technology Kharagpur. Understanding things to the fundamental level leads to new discoveries which lead to advancement in technology. He is passionate about understanding the nature fundamentally with the help of tools like mathematical models, ML models and AI.






Source link

22Apr

This AI Paper from Peking University and Microsoft Proposes LongEmbed to Extend NLP Context Windows


Embedding models are fundamental tools in natural language processing (NLP), providing the backbone for applications like information retrieval and retrieval-augmented generation. These models transform the text into a numerical format that machines can process, which is crucial for understanding and manipulating language. Traditionally, these models are restricted by a narrow context window, typically handling no more than 512 tokens. This limitation restricts their use in scenarios demanding the analysis of extended documents, such as legal contracts or detailed academic reviews.

Existing research in NLP embedding models has progressively focused on extending context capabilities. Early models like BERT utilized absolute position embedding (APE), while more recent innovations like RoFormer and LLaMA incorporate rotary position embedding (RoPE) for handling longer texts. Notable models such as Longformer and BigBird leverage sparse attention mechanisms to process extended documents efficiently. These advancements underscore the evolution from traditional embeddings to sophisticated models capable of managing significantly larger sequences, enhancing the applicability of NLP across various complex and lengthy text processing scenarios.

Researchers from Peking University and Microsoft have proposed LongEmbed, a method to extend the context window of embedding models up to 32,000 tokens without additional training. This method uniquely employs position interpolation and RoPE, differentiating it by its capacity to efficiently manage significantly larger text sequences while maintaining the model’s baseline performance on shorter inputs.

Specifically, the methodology detailed in the study centers around two primary strategies: position interpolation and rotary position embedding (RoPE). These techniques are applied to existing models, notably E5Base and GTEBase, to extend their context-handling capabilities. The position interpolation method extends the models’ original context window by linearly interpolating existing position embeddings. Meanwhile, RoPE is implemented to enhance the scalability of handling longer sequences. The effectiveness of these methods is evaluated on the LongEmbed benchmark, specifically designed for this research, and includes both synthetic and real-world tasks aimed at testing extended context capabilities across diverse document lengths.

The benchmarking results from the LongEmbed framework indicate significant improvements in model performance. Models utilizing the extended context window demonstrated a 20% increase in retrieval accuracy on documents exceeding 4,000 tokens compared to their standard configurations. Moreover, models enhanced with RoPE saw an average accuracy gain of 15% across all tested document lengths. These quantitative findings confirm that the applied methodologies preserve the original model efficiencies for shorter texts and substantially improve their applicability and precision for extended text sequences.

To conclude, the research introduced LongEmbed, a method that significantly extends the context window of NLP embedding models without requiring retraining. By integrating position interpolation and rotary position embedding, the research successfully expands model capacities to process texts up to 32,000 tokens, enhancing retrieval accuracy and applicability in real-world scenarios. The effectiveness of these methods is validated through comprehensive benchmark testing, confirming that these innovations enable existing models to handle extended texts efficiently, making them more versatile and applicable to a broader range of tasks.


Check out the Paper and GithubAll credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter. Join our Telegram Channel, Discord Channel, and LinkedIn Group.

If you like our work, you will love our newsletter..

Don’t Forget to join our 40k+ ML SubReddit


Nikhil is an intern consultant at Marktechpost. He is pursuing an integrated dual degree in Materials at the Indian Institute of Technology, Kharagpur. Nikhil is an AI/ML enthusiast who is always researching applications in fields like biomaterials and biomedical science. With a strong background in Material Science, he is exploring new advancements and creating opportunities to contribute.






Source link

21Apr

Comparative Analysis of Top 14 Vector Databases: Features, Performance, and Scalability Insights


Vector databases have become increasingly prominent, especially in applications that involve machine learning, image processing, and similarity searches. Unlike traditional databases that store data as scalar values (numbers and strings), vector databases are designed to handle multidimensional data points, typically represented as vectors. These vectors can be used to model complex items like images, videos, and text in a format that machines can interpret for tasks such as content recommendation, anomaly detection, and more. Let’s explore 14 different vector databases and provide a comparative analysis of several key parameters. 

Faiss, developed by Facebook AI, is designed for efficient similarity search & clustering of dense vectors. It works well with GPUs for maximum efficiency.

  • Pros: High performance, GPU acceleration, robust in handling very large vector sets.
  • Cons: Mainly focused on similarity search, less flexibility for other database operations.

Milvus

An open-source vector database, Milvus is optimized for scalable similarity search and AI applications. It supports multiple metric types and is highly scalable.

  • Pros: Highly scalable, supports multiple metrics, easy integration with AI frameworks.
  • Cons: Requires a good understanding of its architecture for optimal setup.

Annoy (Approximate Nearest Neighbors Oh Yeah)

Annoy is a C++ library with Python bindings that searches for points in space that are close to a given query point. It is primarily used for music and image recommendation systems.

  • Pros: Very fast, lightweight, allows for static files.
  • Cons: It is not as scalable for large data sets, such as an in-memory database.

ScaNN (Scalable Nearest Neighbors)

Developed by Google, ScaNN is a library designed to search for nearest neighbors in a large dataset efficiently. It works well with TensorFlow.

  • Pros: High performance, integrates well with TensorFlow, efficient on large datasets.
  • Cons: Complexity in setup and tuning.

Hnswlib

A user-friendly library that enables efficient and fast approximate nearest neighbor search. It is based on the Hierarchical Navigable Small World (HNSW) graph.

  • Pros: Fast search times, efficient memory usage, and open-source.
  • Cons: Limited by the characteristics of the HNSW algorithm, more suitable for academic use.

Pinecone

A fully managed vector database service that simplifies building and scaling vector search applications. It provides an easy-to-use API.

  • Pros: Managed service, easy scaling, intuitive API.
  • Cons: Cost can be a factor as it is a managed service with less control over the underlying hardware.

Weaviate

An open-source smart vector search engine that supports GraphQL and RESTful APIs. It includes features like automatic machine learning indexing.

  • Pros: Feature-rich, supports semantic search, integrated ML capabilities.
  • Cons: Requires resources for optimal operation complex configuration.

Qdrant

Qdrant is a vector search engine that supports persistent storage and performs well. It focuses on maintaining the balance between search speed and update speed.

  • Pros: Balances search and update speeds, persistent storage, and good documentation.
  • Cons: Relatively new, smaller community.

Vespa

Developed by Yahoo, Vespa is an engine for low-latency computation over large data sets. It’s highly scalable and supports machine-learned model inference.

  • Pros: High scalability, built-in machine learning support, comprehensive features.
  • Cons: Complex architecture, steeper learning curve.

Vald

A highly scalable distributed vector database that uses Kubernetes. Vald offers automatic indexing and backup features.

  • Pros: Kubernetes native, automatic indexing, resilient design.
  • Cons: Complexity of deployment requires Kubernetes knowledge.

Vectorflow

Vectorflow is a vector database designed for real-time vector indexing and search in a distributed environment.

  • Pros: Real-time operations support distributed architecture.
  • Cons: It needs to be known, and there may be a smaller support community.

Jina

An open-source neural search framework that provides cloud-native neural search solutions powered by AI and deep learning.

  • Pros: AI-driven, supports deep learning models, and is highly extensible.
  • Cons: It can be overkill for simpler search tasks and requires deep learning expertise.

Elasticsearch with vector plugins

Elasticsearch is a broadly used search engine that can effectively handle vector data when equipped with vector search plugins.

  • Pros: Extensive community, robust features, well-documented.
  • Cons: Plugins required for vector functionality can be resource-intensive.

Zilliz

A cloud-native vector database designed for AI and big data challenges. It leverages the power of modern GPUs for processing.

  • Pros: GPU acceleration, designed for AI applications, scalable.
  • Cons: GPU dependency might increase costs, and it is relatively new.

Comparative Table

To better compare the vector databases, let’s break down the parameters into more specific categories and check each database’s capabilities, such as particular features, technology compatibility, and operational nuances.

Comparative Table: Different Vector Databases

In conclusion, the landscape of vector databases is rich and varied, with each platform offering unique strengths tailored to specific use cases and technical requirements. From highly scalable solutions like Milvus and Elasticsearch, designed to handle enormous datasets and complex queries, to specialized offerings like Faiss and Annoy, optimized for speed and efficiency in similarity searches, there is a vector database to suit nearly any need. Managed services like Pinecone are easy and simple, making them ideal for those seeking quick deployment without deep technical overhead. Meanwhile, platforms like Vespa and Jina bring advanced capabilities like real-time indexing and deep learning integration, which are suitable for cutting-edge AI applications. Choosing the right vector database requires careful consideration of scalability, performance, ease of use, and feature set, as highlighted in the detailed comparison table.


Hello, My name is Adnan Hassan. I am a consulting intern at Marktechpost and soon to be a management trainee at American Express. I am currently pursuing a dual degree at the Indian Institute of Technology, Kharagpur. I am passionate about technology and want to create new products that make a difference.




Source link

20Apr

Meta Launches Llama-3 Powered Meta AI Chatbot Assistant to Compete with ChatGPT


Meta has officially introduced its new AI assistant, an AI chatbot called Meta AI, powered by Meta’s latest and most capable openly available LLM, Meta Llama 3. Since the big bang in the popularity of AI chatbots with OpenAI’s ChatGPT, almost every major organization wants to get involved, from Google with Gemini to Meta with probably the most capable AI chatbot currently, Meta AI powered by Llama 3. 

What is Llama 3: 

Before starting with Meta AI, you must know what Llama 3 is, which powers the chatbot, and why Meta calls it the most capable openly available large language model (LLM). Meta’s Llama 3 is an impressive LLM that has outperformed its previous version, Llama 2, in terms of performance. With a staggering 8 to 70 billion parameters, Llama 3 surpasses any other model in its class. The model’s improved pre-training and post-training processes have significantly improved various tasks.

Llama 3 reduces errors, enhances response diversity, and provides better alignment. Meta has also developed a new human evaluation set covering 12 use cases to ensure real-world performance, like asking for advice, brainstorming, classification, closed question answering, coding, creative writing, extraction, inhabiting a character/persona, open question answering, reasoning, rewriting, and summarization. 

What is Meta AI Chatbot:

Coming back to the main event, Meta AI, an advanced AI assistant built with Meta Llama 3, is free to use on your phone through apps like Facebook, Instagram, WhatsApp, and Messenger. You can use Meta AI to accomplish various tasks and stay connected with the things that matter most to you.

With its increasing popularity, it is gaining global attraction, and more people worldwide can interact with it in more ways than ever before. Earlier, you could use Meta AI only on Facebook, Instagram, WhatsApp, and Messenger, but now Meta AI has its own standalone website to get things done right from your computer.

Using Meta AI on Instagram:

You can find Meta AI in the search panel, in a blue ring shape, or you can chat with Meta AI directly in direct messages (DMs). 

You can search for anything without leaving the Meta app you are in.

You can generate AI images directly by prompting /Imagine and the image prompt. You can even animate it with a single click. 

You can do the same thing on Messenger and WhatsApp.

Meta AI on Computer/Website:

You should visit the Meta AI website

From here, it is a simple chat interface you can interact with and ask Meta AI anything. Meta AI is limitless.

You can generate AI images for free; feature OpenAI charges $20 to access.

However, Meta AI is only publicly available in 13 countries (Australia, Canada, Ghana, Jamaica, Malawi, New Zealand, Nigeria, Pakistan, Singapore, South Africa, Uganda, Zambia, and Zimbabwe). If you are not from these countries, you must use a VPN to access Meta AI properly.

Making on Meta AI

Meta built Meta AI on top of Llama 3 and took additional measures to ensure responsible use. Meta’s goal was to provide a safe and helpful assistant for free within Meta’s apps. Meta improved Meta AI’s responses to people’s prompts and questions and taught it specific instructions and responses to make it more helpful. Meta evaluated Meta AI’s performance against benchmarks and applied safeguards at the prompt and response level. Meta also built feedback tools within Meta AI for ongoing model training and improvement. Meta seeks transparency and lets users know that Meta AI is AI technology for everyone.

In Conclusion:

Meta AI is an intelligent assistant that helps you expand your knowledge, get things done, create, and connect. You can use Meta AI to research topics, explore interests, get advice, and learn new hobbies. You can even get inspired and visualize your ideas with Meta’s latest image-generation technology. Experience improved social connections by making plans, sparking conversations, and giving recommendations. You can use Meta AI in any of Meta’s apps or get started at Meta AI on the web.


Nishant is a growth product manager at Marktechpost. Nishant is involved in crafting innovative strategies to enhance user engagement and drive product expansion. With a keen eye for analytics and a passion for technology, he navigates the dynamic intersection of marketing and technology, propelling the company towards sustainable growth and market leadership.




Source link

20Apr

This AI Paper from CMU Introduces AgentKit: A Machine Learning Framework for Building AI Agents Using Natural Language


Agent-based systems in Artificial Intelligence are ones where AI agents perform tasks autonomously within digital environments. Developing intelligent agents that can understand complex instructions and interact dynamically with their environment poses a significant technological challenge. A prevalent issue in agent design is the reliance on sophisticated programming techniques. Traditionally, agents are constructed using code-intensive methods, necessitating a deep familiarity with specific APIs and often restricting flexibility. Such approaches can stifle innovation and accessibility, limiting the potential applications of AI agents outside specialized domains.

Existing research includes the integration of LLMs like GPT-4 and Chain-of-Thought prompting in agent systems for enhanced planning and interaction. Frameworks like LangChain have refined agent operations, enabling more responsive task management. Innovations by researchers have applied these models to complex scenarios like open-world gaming, using structured prompting to guide agent behavior effectively. These models and frameworks demonstrate a significant shift towards more adaptable and intuitive AI architectures, facilitating dynamic responses and detailed task execution in varying environments.

In a collaborative effort, researchers from Carnegie Mellon University, NVIDIA, Microsoft, and Boston University have introduced AgentKit, a framework enabling users to construct AI agents using natural language instead of code. This method is distinct because it employs a graph-based design where each node represents a sub-task defined by language prompts. This structure allows complex agent behaviors to be pieced together intuitively, enhancing user accessibility and system flexibility.

AgentKit employs a structured methodology, mapping each task to a directed acyclic graph (DAG) node. These nodes, representing individual tasks, are interconnected based on task dependencies, ensuring logical progression and systematic execution. As mentioned, the nodes utilize LLMs, specifically GPT-4, to interpret and generate responses to natural language prompts. The framework dynamically adjusts these nodes during execution, allowing real-time response to environmental changes or task demands. Each node’s output is fed into subsequent nodes, maintaining a continuous and efficient workflow. The methodology is geared towards both flexibility in task management and precision in executing complex sequences of operations.

In testing, AgentKit significantly enhanced task efficiency and adaptability. For instance, the Crafter game simulation improved task completion by 80% compared to existing methods. In the WebShop scenario, AgentKit achieved a 5% higher performance than state-of-the-art models, showcasing its effectiveness in real-time decision-making environments. These results confirm AgentKit’s capability to manage complex tasks through intuitive setups. They illustrate its practical applicability across diverse application domains, achieving robust and measurable improvements in agent-based task execution.

To conclude, AgentKit represents a significant advancement in AI agent development, simplifying the creation of complex agents through natural language prompts instead of traditional coding. By integrating a graph-based design with large language models like GPT-4, AgentKit allows users to dynamically construct and modify AI behaviors. The framework’s successful application in diverse scenarios, such as gaming and e-commerce, demonstrates its effectiveness and versatility. This research highlights the potential for broader adoption of intuitive, accessible AI technologies in various industries.


Check out the Paper and GithubAll credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter. Join our Telegram Channel, Discord Channel, and LinkedIn Group.

If you like our work, you will love our newsletter..

Don’t Forget to join our 40k+ ML SubReddit


For Content Partnership, Please Fill Out This Form Here..


Nikhil is an intern consultant at Marktechpost. He is pursuing an integrated dual degree in Materials at the Indian Institute of Technology, Kharagpur. Nikhil is an AI/ML enthusiast who is always researching applications in fields like biomaterials and biomedical science. With a strong background in Material Science, he is exploring new advancements and creating opportunities to contribute.






Source link

19Apr

HR Compliance Program Manager at LE001 San Juan


Love what you do; Love where you live

Reporting directly to the CHRO, this position provides advice and consultation on complex topics and regulations, conducts comprehensive investigations of complex and/or multiple claims, prepares narrative reports, and implements and monitors programs to ensure compliance with federal and state regulations, as well as SJRMC policies and procedures.

Required Behaviors:

  • As you go about fulfilling this mission, your work habits and work relationships should embody SJRMC’s values. These values are our culture, our identity as an organization. Sacred Trust, Personal Reverence, Thoughtful Anticipation, Team Accountability and Creative Vitality ask more of us than merely completing some list of tasks. Our values ask for a deeper level of commitment, and what is asked of us we freely give because we believe in our mission.

Required Qualifications:

  • Bachelor’s Degree in Business Administration, Health Care Management, or Public Policy
  • Five (5) years of experience directly related to the duties and responsibilities specified
  • Completed degree(s) from an accredited institution that are above the education requirements may be substituted for experience on a year for year basis
  • Proficiency in Microsoft Office Suite applications, including Outlook, Word, Excel, and PowerPoint
  • Effective writing skills with the ability to communicate key messages in an easy-to-understand way
  • Ability to think, plan, and execute autonomously
  • Experience with planning, prioritization, and attention to detail
  • Efficient collaboration and listening skills; building upon a wide network, sparking great ideas, gathering, and incorporating feedback from multiple people and groups

Preferred Qualifications:

  • MBA, MPH, or MPP in Business Administration, Health Care Management, or Public Policy
  • Five (5) years of project/program management experience; able to manage multiple projects and deliverables simultaneously across workgroups and stakeholders
  • Five (5) years of experience in change management and process improvement techniques; ability to motivate yourself and others to take on new challenges
  • Knowledge of accrediting/regulatory body requirements

Duties and Responsibilities:

  • Creates and leads HR Compliance Programs that are linked through common departmental or organizational goals and strategies including those related to Licensure and Certification
  • Partners with SJRMC Compliance to ensure perpetual survey-readiness
  • Partners with HRIS and Learning & Development to build and maintain HR Compliance Technology Infrastructure
  • Partners with Talent Acquisition and HR Operations to ensure onboarding workflow and technology reliably gathers all new hire record requirements
  • Partners with Learning & Development and HR Operations leadership to optimize learning assignments that meet regulatory requirements
  • Partners with HR Business Partners to manage ER incidents
  • Performs internal investigations
  • Performs Exit Interviews (in concert with Stay Interviews)
  • Articulates vision, goals, and objectives for all HR Compliance Programs (i.e., manages strategic projects and ensures deliverables are aligned with the organization’s strategy)
  • Manages the execution of strategies through program design and implementation, ensuring strategic alignment with the Senior Leadership Team
  • Partners with stakeholders to build cross functional strategy and implementation teams; facilitates identification of program scope, manages associated project implementation, communicates progress to HR Leadership Team and Senior Leadership Team, and drives timely execution to obtain desired results
  • Identify and manage risks associated with individual projects and overall program development and implement mitigation strategies to remove barriers
  • Develops effective communication strategies to facilitate initiative rollout
  • Ensures change management feasibility, plan, and implement accordingly
  • Develops collaborative relationships with key stakeholders and other consulting resources to help drive and lead cross functional project efforts
  • Updates key stakeholders of initiative progress, challenges, and barriers
  • Develops and enforces policies and procedures for operational efficiency and quality
  • Facilitates the Senior Leadership Team’s decision making by analyzing strategic alternatives and developing recommendations
  • Escalates Compliance Program issues, risks, or other appropriate information to HR Leadership Team or Senior Leadership Teams as appropriate
  • Manages multiple projects and assignments; assemble project teams, collaborate on responsibilities, develop project schedules, and determine resources needed to achieve milestones and overall objectives/outcomes
  • Collaborates with the Compliance Department to implement transparent policies and procedures to define compliance controls within CWE Departments
  • Provides timely and accurate guidance to all CWE Departments on compliance matters; to ensure compliance with all policies and procedures
  • Evaluates and improves CWE systems, processes, and interactions between CWE and SJRMC Departments
  • Acts as a liaison between CWE and regulatory agencies, ensuring seamless communication and response to external audits and/or investigations involving regulatory and compliance matters
  • Contributes to project teams, assigning and monitoring work, and providing guidance and leadership
  • Educates and collaborates with CWE Departments to identify innovation opportunities and adopts improvement actions
  • Analyzes compliance data to identify trends and risks; periodically reports to HR Leadership Team on metrics and trends
  • Responsible for researching best practices and creating and/or revising HR Compliance policies and procedures in conjunction with key stakeholders; to align with regulatory requirements, industry benchmarks, and programmatic functions
  • Participates in workgroups and provides advice related to HR Compliance requirements
  • Supports the CHRO in preparing reports and updates for the Board of Directors, Senior Leadership Team, SJRMC/SJHP Leadership Team, and HR Leadership Team
  • Manages all San Juan Regional Medical Center (SJRMC) and San Juan Health Partners (SJHP) federal and state labor law posters (including all NM and CO clinics)

Physical Demands and Environmental Work Conditions:

  • Prolonged periods of sitting at a desk and working on a computer
  • Occasional irregular work hours (to meet the needs of SJRMC)
  • Must be able to lift twenty-five (25) pounds
  • Occasional: lifting, carrying, reaching, bending, pushing, pulling, and stooping
  • Must be able to: talk (ordinary and conversations), hear (ordinary and conversations), and see (acuity: near and far, depth perception, and color vision)



Source link

19Apr

Megalodon: A Deep Learning Architecture for Efficient Sequence Modeling with Unlimited Context Length


Developing and enhancing models capable of efficiently managing extensive sequential data is paramount in modern computational fields. This necessity is particularly critical in natural language processing, where models must process long text streams seamlessly, retaining context without compromising processing speed or accuracy. One of the key challenges within this scope is the traditional reliance on Transformer architectures, which, despite their broad adoption, suffer from quadratic computational complexity. 

Existing research includes the Transformer architecture, which, despite its efficacy, suffers from high computational costs with longer sequences. Alternatives like linear attention mechanisms and state space models have been developed to reduce this cost, though often at the expense of performance. With its gated attention mechanism and exponential moving average, the LLAMA model and the MEGA architecture aim to address these limitations. However, these models still face challenges in scaling and efficiency, particularly in large-scale pretraining and handling extended data sequences.

Researchers from Meta, the University of Southern California, Carnegie Mellon University, and the University of California San Diego have introduced MEGALODON, a model designed to efficiently handle sequences of unlimited length—a capability that existing models struggle with. By integrating a Complex Exponential Moving Average (CEMA) and timestep normalization, MEGALODON offers reduced computational load and improved scalability, distinguishing itself from traditional Transformer models exhibiting quadratic computational growth with sequence length.

MEGALODON employs a combination of CEMA, timestep normalization, and a normalized attention mechanism. These technical components are crucial for modeling long sequences with high efficiency and low memory cost. The model has been rigorously tested on various language processing benchmarks, including multi-turn conversations, long-document comprehension, and extensive language modeling tasks. MEGALODON was benchmarked against datasets specifically designed for long-context scenarios, such as the Scrolls dataset for long-context QA tasks and PG19, which consists of long literary texts to demonstrate its efficacy and versatility. 

MEGALODON demonstrated quantifiable improvements in performance metrics. It recorded a training loss of 1.70, positioned between LLAMA2-7B, which registered a loss of 1.75, and LLAMA2-13B at 1.67. Regarding specific benchmarks, MEGALODON outperformed a standard Transformer model by achieving a lower perplexity rate on the Scrolls dataset, measuring at 23, compared to the Transformer’s 30. These results affirm MEGALODON‘s advanced processing capabilities for lengthy sequential data, substantiating its efficiency and effectiveness across varied linguistic tasks.

To conclude, the MEGALODON model marks a significant advancement in sequence modeling, addressing the inefficiencies of traditional Transformer architectures with innovative approaches like CEMA and timestep normalization. By achieving a training loss of 1.70 and demonstrating improved performance on challenging benchmarks such as the Scrolls dataset, MEGALODON proves its capability to handle extensive sequences effectively. This research enhances the processing of long data sequences and sets a new standard for future developments in natural language processing and related fields.


Check out the PaperAll credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter. Join our Telegram Channel, Discord Channel, and LinkedIn Group.

If you like our work, you will love our newsletter..

Don’t Forget to join our 40k+ ML SubReddit


For Content Partnership, Please Fill Out This Form Here..


Nikhil is an intern consultant at Marktechpost. He is pursuing an integrated dual degree in Materials at the Indian Institute of Technology, Kharagpur. Nikhil is an AI/ML enthusiast who is always researching applications in fields like biomaterials and biomedical science. With a strong background in Material Science, he is exploring new advancements and creating opportunities to contribute.






Source link

18Apr

Hugging Face Researchers Introduce Idefics2: A Powerful 8B Vision-Language Model Elevating Multimodal AI Through Advanced OCR and Native Resolution Techniques


As digital interactions become increasingly complex, the demand for sophisticated analytical tools to understand and process this diverse data intensifies. The core challenge involves integrating distinct data types, primarily images, and text, to create models that can effectively interpret and respond to multimodal inputs. This challenge is critical for applications ranging from automated content generation to enhanced interactive systems.

Existing research includes models like LLaVa-NeXT and MM1, which are known for their robust multimodal capabilities. The LLaVa-NeXT series, particularly the 34B variant, and MM1-Chat models have set benchmarks in visual question answering and image-text integration. Gemini models like Gemini 1.0 Pro further push performance in complex AI tasks. DeepSeek-VL specializes in visual question answering, while Claude 3 Haiku excels in generating narrative content from visual inputs, showcasing diverse approaches to blending visual and textual data within AI frameworks.

Hugging Face Researchers have introduced Idefics2, a powerful 8B parameter vision-language model designed to enhance the integration of text and image processing within a single framework. This method contrasts with previous models, which often required the resizing of images to fixed dimensions, potentially compromising the detail and quality of visual data. This capability, derived from the NaViT strategy, enables Idefics2 to process visual information more accurately and efficiently. Integrating visual features into the language backbone via learned Perceiver pooling and an MLP modality projection further distinguishes this model, facilitating a deeper and more nuanced understanding of multimodal inputs.

The model was pre-trained on a blend of publicly available resources, including Interleaved web documents, image-caption pairs from the Public Multimodal Dataset and LAION-COCO, and specialized OCR data from PDFA, IDL, and Rendered-text. Moreover, Idefics2 was fine-tuned using “The Cauldron,” a carefully curated compilation of 50 vision-language datasets. This fine-tuning phase employed technologies like Lora for adaptive learning and specific fine-tuning strategies for newly initialized parameters in the modality connector, which underpins the distinct functionalities of its various versions—ranging from the generalist base model to the conversationally adept Idefics2-8B-Chatty, poised for release. Each version is designed to excel in different scenarios, from basic multimodal tasks to complex, long-duration interactions.

Versions of Idefics2:

Idefics2-8B-Base:

This version serves as the foundation of the Idefics2 series. It has 8 billion parameters and is designed to handle general multimodal tasks. The base model is pre-trained on a diverse dataset, including web documents, image-caption pairs, and OCR data, making it robust for many basic vision-language tasks.

Idefics2-8B:

The Idefics2-8B extends the base model by incorporating fine-tuning on ‘The Cauldron,’ a specially prepared dataset consisting of 50 manually curated multimodal datasets and text-only instruction fine-tuning datasets. This version is tailored to perform better on complex instruction-following tasks, enhancing its ability to understand and process multimodal inputs more effectively.

Idefics2-8B-Chatty (Coming Soon):

Anticipated as an advancement over the existing models, the Idefics2-8B-Chatty is designed for long conversations and deeper contextual understanding. It is further fine-tuned for dialogue applications, making it ideal for scenarios that require extended interactions, such as customer service bots or interactive storytelling applications.

Improvements over Idefics1:

  • Idefics2 utilizes the NaViT strategy for processing images in native resolutions, enhancing visual data integrity.
  • Enhanced OCR capabilities through specialized data integration improve text transcription accuracy.
  • Simplified architecture using vision encoder and Perceiver pooling boosts performance significantly over Idefics1.

In testing, Idefics2 demonstrated exceptional performance across multiple benchmarks. The model achieved an 81.2% accuracy in Visual Question Answering (VQA) on standard benchmarks, significantly surpassing its predecessor, Idefics1. Furthermore, Idefics2 showed a 20% improvement in character recognition accuracy in document-based OCR tasks compared to earlier models. The enhancements in OCR capabilities specifically reduced the error rate from 5.6% to 3.2%, establishing its efficacy in practical applications requiring high levels of accuracy in text extraction and interpretation.

To conclude, the research introduced Idefics2, a visionary vision-language model that integrates native image resolution processing and advanced OCR capabilities. The model demonstrates significant advancements in multimodal AI, achieving top-tier results in visual question answering and text extraction tasks. By maintaining the integrity of visual data and enhancing text recognition accuracy, Idefics2 represents a substantial leap forward, promising to facilitate more accurate and efficient AI applications in fields requiring sophisticated multimodal analysis.


Check out the HF Project Page and Blog. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter. Join our Telegram Channel, Discord Channel, and LinkedIn Group.

If you like our work, you will love our newsletter..

Don’t Forget to join our 40k+ ML SubReddit


For Content Partnership, Please Fill Out This Form Here..


Nikhil is an intern consultant at Marktechpost. He is pursuing an integrated dual degree in Materials at the Indian Institute of Technology, Kharagpur. Nikhil is an AI/ML enthusiast who is always researching applications in fields like biomaterials and biomedical science. With a strong background in Material Science, he is exploring new advancements and creating opportunities to contribute.






Source link

17Apr

Dataset Reset Policy Optimization (DR-PO): A Machine Learning Algorithm that Exploits a Generative Model’s Ability to Reset from Offline Data to Enhance RLHF from Preference-based Feedback


Reinforcement Learning (RL) continuously evolves as researchers explore methods to refine algorithms that learn from human feedback. This domain of learning algorithms deals with challenges in defining and optimizing reward functions critical for training models to perform various tasks ranging from gaming to language processing.

A prevalent issue in this area is the inefficient use of pre-collected datasets of human preferences, often overlooked in the RL training processes. Traditionally, these models are trained from scratch, ignoring existing datasets’ rich, informative content. This disconnect leads to inefficiencies and a lack of utilization of valuable, pre-existing knowledge. Recent advancements have introduced innovative methods that effectively integrate offline data into the RL training process to address this inefficiency.

Researchers from Cornell University, Princeton University, and Microsoft Research introduced a new algorithm, the Dataset Reset Policy Optimization (DR-PO) method. This method ingeniously incorporates preexisting data into the model training rule and is distinguished by its ability to reset directly to specific states from an offline dataset during policy optimization. It contrasts with traditional methods that begin every training episode from a generic initial state.

The DR-PO method enhances offline data by allowing the model to ‘reset’ to specific, beneficial states already identified as useful in the offline data. This process reflects real-world conditions where scenarios are not always initiated from scratch but are often influenced by prior events or states. By leveraging this data, DR-PO improves the efficiency of the learning process and broadens the application scope of the trained models.

DR-PO employs a hybrid strategy that blends online and offline data streams. This method capitalizes on the informative nature of the offline dataset by resetting the policy optimizer to states previously identified as valuable by human labelers. The integration of this method has demonstrated promising improvements over traditional techniques, which often disregard the potential insights available in pre-collected data.

DR-PO has shown outstanding results in studies involving tasks like TL;DR summarization and the Anthropic Helpful Harmful dataset. DR-PO has outperformed established methods like Proximal Policy Optimization (PPO) and Direction Preference Optimization (DPO). In the TL;DR summarization task, DR-PO achieved a higher GPT4 win rate, enhancing the quality of generated summaries. In head-to-head comparisons, DR-PO’s approach to integrating resets and offline data has consistently demonstrated superior performance metrics.

In conclusion, DR-PO presents a significant breakthrough in RL. DR-PO overcomes traditional inefficiencies by integrating pre-collected, human-preferred data into the RL training process. This method enhances learning efficiency by utilizing resets to specific states identified in offline datasets. Empirical evidence demonstrates that DR-PO surpasses conventional approaches such as Proximal Policy Optimization and Direction Preference Optimization in real-world applications like TL;DR summarization, achieving superior GPT4 win rates. This innovative approach streamlines the training process and maximizes the utility of existing human feedback, setting a new benchmark in adapting offline data for model optimization.


Check out the Paper and GithubAll credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter. Join our Telegram Channel, Discord Channel, and LinkedIn Group.

If you like our work, you will love our newsletter..

Don’t Forget to join our 40k+ ML SubReddit


Want to get in front of 1.5 Million AI Audience? Work with us here


Hello, My name is Adnan Hassan. I am a consulting intern at Marktechpost and soon to be a management trainee at American Express. I am currently pursuing a dual degree at the Indian Institute of Technology, Kharagpur. I am passionate about technology and want to create new products that make a difference.






Source link

Protected by Security by CleanTalk