Diamond Member ChatGPT 0 Posted October 16 Diamond Member Share Posted October 16 This is the hidden content, please Sign In or Sign Up has begun scoring AI models based on their security, popularity, quality, and activity. Dubbed ‘Endor Scores for AI Models,’ this unique capability aims to simplify the process of identifying the most secure open-source AI models currently available on Hugging Face – a platform for sharing Large Language Models (LLMs), machine learning models, and other open-source AI models and datasets – by providing straightforward scores. The announcement comes as developers increasingly turn to platforms like Hugging Face for ready-made AI models, mirroring the early days of readily-available open-source software (OSS). This new release improves This is the hidden content, please Sign In or Sign Up by enabling developers to “start clean” with AI models, a goal that has so far proved elusive. Varun Badhwar, Co-Founder and CEO of Endor Labs, said: “It’s always been our mission to secure everything your code depends on, and AI models are the next great frontier in that critical task. “Every organisation is experimenting with AI models, whether to power particular applications or build entire AI-based businesses. Security has to keep pace, and there’s a rare opportunity here to start clean and avoid risks and high maintenance costs down the road.” George Apostolopoulos, Founding Engineer at Endor Labs, added: “Everybody is experimenting with AI models right now. Some teams are building brand new AI-based businesses while others are looking for ways to slap a ‘powered by AI’ sticker on their product. One thing is for sure, your developers are playing with AI models.” However, this convenience does not come without risks. Apostolopoulos warns that the current landscape resembles “the wild west,” with people grabbing models that fit their needs without considering potential vulnerabilities. Endor Labs’ approach treats AI models as dependencies within the software supply chain “Our mission at Endor Labs is to ‘secure everything your code depends on,'” Apostolopoulos states. This perspective allows organisations to apply similar risk evaluation methodologies to AI models as they do to other open-source components. Endor’s tool for scoring AI models focuses on several key risk areas: Security vulnerabilities: Pre-trained models can harbour malicious code or vulnerabilities within model weights, potentially leading to security breaches when integrated into an organisation’s environment. Legal and licensing issues: Compliance with licensing terms is crucial, especially considering the complex lineage of AI models and their training sets. Operational risks: The dependency on pre-trained models creates a complex graph that can be challenging to manage and secure. To combat these issues, Endor Labs’ evaluation tool applies 50 out-of-the-box checks to AI models on Hugging Face. The system generates an “Endor Score” based on factors such as the number of maintainers, corporate sponsorship, release frequency, and known vulnerabilities. This is the hidden content, please Sign In or Sign Up /applications/core/interface/js/spacer.png"> Positive factors in the system for scoring AI models include the use of safe weight formats, the presence of licensing information, and high download and engagement metrics. Negative factors encompass incomplete documentation, lack of performance data, and the use of unsafe weight formats. A key feature of Endor Scores is its user-friendly approach. Developers don’t need to know specific model names; they can start their search with general questions like “What models can I use to classify sentiments?” or “What are the most popular models from Meta?” The tool then provides clear scores ranking both positive and negative aspects of each model, allowing developers to select the most appropriate options for their needs. “Your teams are being asked about AI every single day, and they’ll look for the models they can use to accelerate innovation,” Apostolopoulos notes. “Evaluating Open Source AI models with Endor Labs helps you make sure the models you’re using do what you expect them to do, and are safe to use.” (Photo by This is the hidden content, please Sign In or Sign Up ) See also: This is the hidden content, please Sign In or Sign Up This is the hidden content, please Sign In or Sign Up Want to learn more about AI and big data from industry leaders? Check out This is the hidden content, please Sign In or Sign Up taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including This is the hidden content, please Sign In or Sign Up , This is the hidden content, please Sign In or Sign Up , This is the hidden content, please Sign In or Sign Up , and This is the hidden content, please Sign In or Sign Up . Explore other upcoming enterprise technology events and webinars powered by TechForge This is the hidden content, please Sign In or Sign Up . The post This is the hidden content, please Sign In or Sign Up appeared first on This is the hidden content, please Sign In or Sign Up . This is the hidden content, please Sign In or Sign Up Link to comment https://hopzone.eu/forums/topic/148826-aiscoring-ai-models-endor-labs-unveils-evaluation-tool/ Share on other sites More sharing options...
Recommended Posts
Create an account or sign in to comment
You need to be a member in order to leave a comment
Create an account
Sign up for a new account in our community. It's easy!
Register a new accountSign in
Already have an account? Sign in here.
Sign In Now