AIDeveloper https://www.webpronews.com/developer/aideveloper/ Breaking News in Tech, Search, Social, & Business Sun, 08 Sep 2024 20:40:54 +0000 en-US hourly 1 https://wordpress.org/?v=6.6.2 https://i0.wp.com/www.webpronews.com/wp-content/uploads/2020/03/cropped-wpn_siteidentity-7.png?fit=32%2C32&ssl=1 AIDeveloper https://www.webpronews.com/developer/aideveloper/ 32 32 138578674 Reflection 70B Outperforms GPT-4o: The Rise of Open-Source AI for Developers https://www.webpronews.com/reflection-70b-outperforms-gpt-4o-the-rise-of-open-source-ai-for-developers/ Sun, 08 Sep 2024 20:39:42 +0000 https://www.webpronews.com/?p=607660 The race between open-source models and proprietary systems has hit a turning point in AI development. Reflection 70B, an open-source model, has managed to surpass some of the most powerful models on the market, including GPT-4o, in a variety of benchmarks. Developed by Matt Shumer and a small team at GlaiveAI, Reflection 70B introduces a new era of AI with its unique Reflection-Tuning approach, allowing the model to fix its own mistakes in real-time. For developers, engineers, and tech professionals, the implications of this breakthrough go far beyond a simple improvement in accuracy—it signals a potential paradigm shift in how large language models (LLMs) are built, deployed, and scaled.

Why Reflection 70B Is a Game-Changer

Reflection 70B is not just another LLM in the crowded AI landscape. It’s built using Reflection-Tuning, a technique that enables the model to self-assess and correct its responses during the generation process. Traditionally, models generate an answer and stop there, but Reflection 70B takes things further by employing a post-generation feedback loop. This reflection phase improves the model’s reasoning capabilities and reduces errors, which is especially critical in complex tasks like logic, math, and natural language understanding.

As Shumer explained, “This model is quite fun to use and insanely powerful. With the right prompting, it’s an absolute beast for many use-cases.” This feature allows the model to perform exceptionally well in both zero-shot and few-shot learning environments, beating other state-of-the-art systems like Claude 3.5, Gemini 1.5, and GPT-4o on every major benchmark tested.

Performance on Benchmarks

For AI developers, one of the most compelling reasons to pay attention to Reflection 70B is its performance across a wide range of benchmarks. The model recorded a 99.2% accuracy on the GSM8k benchmark, which is used to evaluate math and logic skills. This score raised eyebrows within the AI community, with many questioning if the model had simply memorized answers. However, independent testers like Jonathan Whitaker debunked this notion by feeding the model problematic questions with incorrect “ground-truth” answers. “I fed the model five questions from GSM8k that had incorrect answers. It got them all right, rather than regurgitating the wrong answers from the dataset,” Whitaker noted, confirming the model’s superior generalization ability.

Shumer emphasizes that the model excels in zero-shot learning, where the AI has to solve problems without any prior examples. In a world where few-shot learning—providing models with several examples before they make predictions—dominates proprietary systems, Reflection 70B stands out for its ability to reason and solve problems with minimal input. “Reflection 70B consistently outperforms other models in zero-shot scenarios, which is crucial for developers working with dynamic, real-world data where examples aren’t always available,” says Shumer.

The Technology Behind Reflection-Tuning

So how exactly does Reflection-Tuning work? The process can be broken down into three key steps: Plan, Execute, Reflect.

  1. Plan: When asked a question, the model first plans how it will tackle the problem, mapping out potential reasoning steps.
  2. Execute: It then executes the plan and generates an initial response based on its reasoning process.
  3. Reflect: Finally, the model pauses, reviews its own answer, and evaluates whether any errors were made. If it finds mistakes, it revises the output before delivering the final response.

This technique mirrors human problem-solving methods, making the model more robust and adaptable to complex tasks. For developers, this approach is especially valuable when dealing with applications that require a high degree of accuracy, such as medical diagnostics, financial forecasting, or legal reasoning. Traditional models might require frequent retraining to achieve comparable results, but Reflection-Tuning enables the model to fine-tune itself on the fly.

In one test, the model was asked to compare two decimal numbers—9.11 and 9.9. Initially, it answered incorrectly but, through its reflection phase, corrected itself and delivered the right answer. This level of introspection is a significant leap forward in AI capabilities and could reduce the need for constant human oversight during AI deployment.

Open-Source Power: Democratizing AI Development

One of the most remarkable aspects of Reflection 70B is that it’s open-source. Unlike proprietary models like GPT-4o or Google’s Gemini, which are locked behind paywalls and closed platforms, Reflection 70B is available to the public. Developers can access the model weights via platforms like Hugging Face, making it easy to integrate and experiment with the model in a variety of applications.

Shumer emphasizes that this open approach has been key to the model’s rapid development. “Just Sahil and I! This was a fun side project for a few weeks,” he explained, highlighting how small teams with the right tools can compete with tech giants. The model was trained with GlaiveAI data, accelerating its capabilities in a fraction of the time it would take larger companies. “Glaive’s data was what took it so far, so quickly,” he added.

This open-access philosophy also allows developers to customize and fine-tune the model for specific use-cases. Whether you’re building a chatbot, automating customer service, or developing a new AI-driven product, Reflection 70B provides a powerful, flexible base.

The 405B Model and Beyond

Reflection 70B isn’t the end of the road for Shumer and his team. They’re already working on the release of Reflection-405B, a larger model that promises even better performance across benchmarks. Shumer is confident that 405B will “outperform Sonnet and GPT-4o by a wide margin.”

The potential applications for this next iteration are vast. Developers can expect Reflection-405B to bring improvements in areas such as multi-modal learning, code generation, and natural language understanding. With the trend toward larger, more complex models, Reflection-405B could become a leading contender in the AI space, challenging not just open-source competitors but proprietary giants as well.

Challenges and Considerations for AI Developers

While the performance of Reflection 70B is undoubtedly impressive, developers should be aware of a few challenges. As with any open-source model, integrating and scaling Reflection 70B for production environments requires a solid understanding of AI infrastructure, including server costs, data management, and security protocols.

Additionally, Reflection-Tuning may introduce latency in applications requiring real-time responses, such as voice assistants or interactive bots. Shumer acknowledges this, noting that the model’s reflection phase can slow down response times, though optimization techniques could mitigate this issue. For developers aiming to use the model in time-sensitive environments, balancing reflection depth and speed will be a key consideration.

An Interesting New Era for Open-Source AI

Reflection 70B is not just an impressive feat of engineering; it’s a sign that open-source models are capable of competing with—and even outperforming—proprietary systems. For AI developers, the model offers a rare combination of accessibility, flexibility, and top-tier performance, all packaged in a framework that encourages community-driven innovation.

As Shumer himself puts it, “This is just the start. I have a few more tricks up my sleeve.” With the release of Reflection-405B on the horizon, developers should be watching closely. The future of AI may no longer be dominated by closed systems, and Reflection 70B has shown that open-source might just be the key to the next breakthrough in AI technology.

]]>
607660
Elon Musk’s xAI Team Brings Colossus 100k H100 Training Cluster Online in Just 122 Days https://www.webpronews.com/elon-musks-xai-team-brings-colossus-100k-h100-training-cluster-online-in-just-122-days/ Mon, 02 Sep 2024 23:22:56 +0000 https://www.webpronews.com/?p=607303 In a move that puts an exclamation point on the massively accelerating pace of artificial intelligence development, Elon Musk announced over the weekend that his xAI team successfully brought the Colossus 100k H100 training cluster online—a feat completed in an astonishing 122 days. This achievement marks the arrival of what Musk is calling “the most powerful AI training system in the world,” with plans to double its capacity in the coming months.

The Birth of Colossus

The Colossus cluster, composed of 100,000 Nvidia H100 GPUs, represents a significant milestone not just for Musk’s xAI but for the AI industry at large. “This is not just another AI cluster; it’s a leap into the future,” Musk tweeted. The system’s scale and speed of deployment are unprecedented, demonstrating the power of a concerted effort between xAI, Nvidia, and a network of partners and suppliers.

Bringing such a massive system online in just 122 days is an accomplishment that has left many industry experts and tech titans in awe. “It’s amazing how fast this was done, and it’s an honor for Dell Technologies to be part of this important AI training system,” said Michael Dell, CEO of Dell Technologies, one of the key partners in the project. The speed and efficiency of this deployment reflect a new standard in AI infrastructure development, one that could reshape the competitive landscape in AI research and application.

A Technological Marvel

The Colossus system is designed to push the boundaries of what AI can achieve. The 100,000 H100 GPUs provide unparalleled computational power, enabling the training of highly complex AI models at speeds that were previously unimaginable. “Colossus isn’t just leading the pack; it’s rewriting what we thought was possible in AI training,” commented xAI’s official ² account, capturing the sentiment of many in the tech community.

The cluster is set to expand even further, with plans to integrate 50,000 H200 GPUs in the near future, effectively doubling its capacity. The H200, Nvidia’s next-generation GPU, is expected to bring enhancements in both performance and energy efficiency, further solidifying Colossus’s position at the forefront of AI development.

Collaboration on a Grand Scale

Colossus’s rapid deployment was made possible by a collaborative effort that included some of the biggest names in technology. Nvidia, Dell, and other partners provided the essential components and expertise necessary to bring this ambitious project to life. The success of Colossus is a testament to the power of collaboration in driving technological innovation.

“Elon Musk and the xAI team have truly outdone themselves,” said Patrick Moorhead, CEO of Moor Insights & Strategy, in response to the announcement. “This project sets a new benchmark for AI infrastructure, and it’s exciting to see what this will enable in terms of AI research and applications.”

Implications for AI Development

The completion of Colossus represents more than just a technical achievement; it has far-reaching implications for the future of AI. With such a powerful system at its disposal, xAI is poised to accelerate the development of advanced AI models, including those that will power applications like autonomous vehicles, robotics, and natural language processing.

The potential of Colossus extends beyond xAI’s immediate goals. As the system scales and evolves, it could become a critical resource for the broader AI community, offering unprecedented capabilities for research and innovation. “This isn’t just innovation; it’s a revolution,” tweeted one xAI supporter, highlighting the broader impact that Colossus could have on the industry.

What’s Next?

As Colossus comes online, the tech world is watching closely to see what comes next. The expansion to 200,000 GPUs is just the beginning, with Musk hinting at even more ambitious plans on the horizon. The speed and scale of this project have set a new standard in the industry, and it’s clear that xAI is not content to rest on its laurels.

For now, the focus will be on leveraging Colossus’s immense power to push the boundaries of AI. Whether it’s through the development of new AI models or the enhancement of existing ones, the possibilities are virtually limitless. As Musk put it, “The future is now, and it’s powered by xAI.”

Congrats to xAI on this massive achievement!

]]>
607303
Llama 3.1: A Massive Upgrade in Open Source AI Technology https://www.webpronews.com/llama-3-1-a-massive-upgrade-in-open-source-ai-technology/ Sun, 01 Sep 2024 16:09:11 +0000 https://www.webpronews.com/?p=607208 In the rapidly evolving landscape of artificial intelligence, Meta’s Llama models have emerged as formidable players, particularly in the open-source domain. The latest iteration, Llama 3.1, represents a significant leap forward, not just in terms of size and capability, but also in its impact on the AI community and industry adoption. With 405 billion parameters, Llama 3.1 is one of the most advanced large language models (LLMs) available today, marking a pivotal moment in the democratization of AI technology.

The Growth and Adoption of Llama

Since its initial release, the Llama series has experienced exponential growth, with downloads surpassing 350 million as of August 2024. This represents a 10x increase from the previous year, underscoring the model’s widespread acceptance and utility across various sectors. Notably, Llama 3.1 alone was downloaded more than 20 million times in just one month, a testament to its growing popularity among developers and enterprises alike.

Meta’s open-source approach with Llama has been instrumental in this rapid adoption. By making the models freely available, Meta has fostered a vibrant ecosystem where innovation thrives. “The success of Llama is made possible through the power of open source,” Meta announced, emphasizing their commitment to sharing cutting-edge AI technology in a responsible manner. This strategy has enabled a wide range of applications, from startups experimenting with new AI solutions to large enterprises integrating Llama into their operations.

Strategic Partnerships and Industry Impact

Llama’s influence extends beyond just the number of downloads. The model’s integration into major cloud platforms like Amazon Web Services (AWS), Microsoft Azure, and Google Cloud has significantly boosted its usage, particularly in enterprise environments. From May to July 2024, Llama’s token volume across these cloud services doubled, and by August, the highest number of unique users on these platforms was for the 405B variant of Llama 3.1. This trend highlights the increasing reliance on Llama for high-performance AI applications.

Industry leaders have been quick to recognize the value that Llama 3.1 brings to the table. Swami Sivasubramanian, VP of AI and Data at AWS, noted, “Customers want access to the latest state-of-the-art models for building AI applications in the cloud, which is why we were the first to offer Llama 2 as a managed API and have continued to work closely with Meta as they released new models.” Similarly, Ali Ghodsi, CEO of Databricks, praised the model’s quality and flexibility, calling Llama 3.1 a “breakthrough for customers wanting to build high-quality AI applications.”

The adoption of Llama 3.1 by enterprises like AT&T, Goldman Sachs, DoorDash, and Accenture further underscores its growing importance. AT&T, for instance, reported a 33% improvement in search-related responses for customer service, attributing this success to the fine-tuning capabilities of Llama models. Accenture is using Llama 3.1 to build custom large language models for ESG reporting, expecting productivity gains of up to 70%.

Technical Advancements in Llama 3.1

The technical prowess of Llama 3.1 is evident in its advanced features and capabilities. The model’s context length has been expanded to 128,000 tokens, enabling it to handle much longer and more complex inputs than previous versions. This makes it particularly effective for tasks like long-form text summarization, multilingual conversational agents, and even complex mathematical reasoning.

Moreover, Llama 3.1 supports eight languages, including English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai, reflecting Meta’s commitment to making AI more accessible globally. The model is also optimized for tool calling, with built-in support for mathematical reasoning and custom JSON functions, making it highly adaptable for a variety of use cases.

The engineering behind Llama 3.1 is as impressive as its features. Meta’s team has meticulously documented the training process, revealing a highly sophisticated approach that balances performance with efficiency. The model was trained on 15 trillion tokens and fine-tuned using over 10 million human-annotated examples, ensuring it performs exceptionally well across a range of tasks.

Open Source and the Future of AI

Meta’s open-source strategy with Llama has not only democratized access to advanced AI models but also set a new standard for transparency and collaboration in the AI community. The release of Llama 3.1, accompanied by a detailed research paper, provides a blueprint for AI developers and researchers to build upon. This move is expected to catalyze further innovation in the field, as developers can now create derivative models and applications with greater ease and lower costs.

Mark Zuckerberg, CEO of Meta, articulated the company’s vision in an open letter, stating, “Open source promotes a more competitive ecosystem that’s good for consumers, good for companies (including Meta), and ultimately good for the world.” This philosophy is already bearing fruit, as evidenced by the creation of over 60,000 derivative models on platforms like Hugging Face.

The open-source nature of Llama 3.1 also addresses some of the ethical concerns surrounding AI development. Meta has integrated robust safety features like Llama Guard 3 and Prompt Guard, designed to prevent data misuse and promote responsible AI deployment. This is particularly crucial as AI systems become more pervasive in industries like finance, healthcare, and customer service.

A Case Study in Open Source Success

One of the most compelling examples of Llama 3.1’s impact is its adoption by Niantic, the company behind the popular AR game Peridot. Niantic integrated Llama to enhance the game’s virtual pets, known as “Dots,” making them more responsive and lifelike. Llama generates each Dot’s reactions in real-time, creating a dynamic and unique experience for players. This use case exemplifies how Llama 3.1 can drive innovation in both consumer and enterprise applications.

Another significant case is Shopify, which uses LLaVA, a derivative of Llama, for product metadata and enrichment. Shopify processes between 40 million to 60 million inferences per day using LLaVA, highlighting the scalability and efficiency of the Llama 3.1 framework.

The Future of AI with Llama

Llama 3.1 is more than just an upgrade; it represents a paradigm shift in how AI models are developed, deployed, and utilized. With its unprecedented scale, performance, and accessibility, Llama 3.1 is poised to become a cornerstone of the AI ecosystem. As more enterprises and developers adopt Llama, the boundaries of what AI can achieve will continue to expand.

The success of Llama 3.1 also reinforces the importance of open-source AI in driving innovation and ensuring that the benefits of AI are widely distributed. As Meta continues to push the envelope with future releases, the AI landscape will undoubtedly become more dynamic, competitive, and inclusive. Whether in academia, industry, or beyond, Llama 3.1 is setting the stage for a new era of AI development.

]]>
607208
Grok-2 Large Beta: A Groundbreaking Leap in AI or Just More Hype? https://www.webpronews.com/grok-2-large-beta-a-groundbreaking-leap-in-ai-or-just-more-hype/ Mon, 26 Aug 2024 13:27:25 +0000 https://www.webpronews.com/?p=606887 The artificial intelligence (AI) landscape has been buzzing with excitement, skepticism, and intrigue since the quiet release of Grok-2 Large Beta, the latest large language model (LLM) from Elon Musk’s xAI. Unlike the typical high-profile launches that accompany such advanced models, Grok-2 slipped onto the scene without a research paper, model card, or academic validation, raising eyebrows across the AI community. But the mystery surrounding its debut has only fueled more interest, prompting many to ask: Is Grok-2 a true revolution in AI, or is it just another iteration in an already crowded field?

A Mysterious Entrance

In a field where transparency and documentation are highly valued, Grok-2’s introduction was unconventional, to say the least. Traditionally, new AI models are accompanied by detailed research papers that explain the model’s architecture, training data, benchmarks, and potential applications. Grok-2, however, arrived with none of these. Instead, it was quietly integrated into a chatbot on Twitter (or X.com), leaving many AI researchers puzzled.

“It’s unusual, almost unheard of, to release a model of this scale without any academic backing or explanation,” remarked an AI researcher. “It raises questions about the model’s capabilities and the motivations behind its release.”

Despite this unconventional launch, Grok-2 quickly demonstrated its potential, performing impressively on several key benchmarks, including the Google Proof Science Q&A Benchmark and the MLU Pro, where it secured a top position, second only to Claude 3.5 Sonic. These early results suggest that Grok-2 could be a serious contender in the LLM space. However, the lack of transparency has led to a mix of curiosity and skepticism within the AI community.

One commenter on the popular ‘AI Explained’ YouTube channel voiced the general sentiment: “No paper? Just a table with benchmarks. What are the performance claims for Grok-2 really based on? Benchmarks have been repeatedly proven meaningless by this point.”

The Scaling Debate: Beyond Just Bigger Models?

One of the most contentious topics in AI is the concept of scaling—expanding a model’s size, data intake, and computational power to enhance its performance. This debate has been reignited by Grok-2’s release, particularly in light of a recent paper from Epoch AI, which predicts that AI models could be scaled up by a factor of 10,000 by 2030. Such a leap could revolutionize the field, potentially bringing us closer to AI that can reason, plan, and interact with humans on a level akin to human cognition.

The Epoch AI paper suggests that scaling could lead to the development of “world models,” where AI systems develop sophisticated internal representations of the world, enabling them to understand and predict complex scenarios better. This could be a significant step toward achieving Artificial General Intelligence (AGI), where AI systems can perform any intellectual task that a human can.

However, this vision is not universally accepted. “We’ve seen time and time again that more data and more parameters don’t automatically lead to more intelligent or useful models,” cautioned an AI critic. “What we need is better data, better training techniques, and more transparency in how these models are built and evaluated.”

This skepticism is echoed by many in the AI field. As another user on the ‘AI Explained’ channel noted, “Does anybody really believe that scaling alone will push transformer-based ML up and over the final ridge before the arrival at the mythical summit that is AGI?” This highlights a broader concern that merely making models larger may not address the fundamental limitations of current AI architectures.

Testing Grok-2: Early Performance and Challenges

In the absence of official documentation, independent AI enthusiasts and researchers have taken it upon themselves to test Grok-2’s capabilities. The Simple Bench project, an independent benchmark designed to test reasoning and problem-solving abilities, has become a key tool in this effort. According to the creator of Simple Bench, who also runs the ‘AI Explained’ channel, Grok-2 has shown promise, though it still has room for improvement.

“Grok-2’s performance was pretty good, mostly in line with the other top models on traditional benchmarks,” the creator shared. “But it’s not just about scores—it’s about how these models handle more complex, real-world tasks.”

Simple Bench focuses on tasks requiring models to understand and navigate cause-and-effect relationships, which are often overlooked by traditional benchmarks. While Grok-2 performed well in many areas, it fell short in tasks where Claude 3.5 Sonic excelled, particularly those that required deeper reasoning and contextual understanding.

Reflecting on the importance of benchmarks like Simple Bench, one commenter observed, “What I like about Simple Bench is that it’s ball-busting. Too many of the recent benchmarks start off at 75-80% on the current models. A bench that last year got 80% and now gets 90% is not as interesting anymore for these kinds of bleeding-edge discussions on progress.” This sentiment underscores the need for benchmarks that challenge AI models to push beyond the easily achievable, testing their limits in more meaningful ways.

The Ethical Dilemmas: Deepfakes and Beyond

As AI models like Grok-2 become more sophisticated, they also introduce new ethical challenges, particularly concerning the generation of highly convincing deepfakes in real-time. With tools like Flux, Grok-2’s image-generating counterpart, the line between reality and digital fabrication is blurring at an alarming rate.

“We’re not far from a world where you won’t be able to trust anything you see online,” warned an AI enthusiast. “The line between reality and fabrication is blurring at an alarming rate.”

The potential for misuse is significant, ranging from spreading misinformation to manipulating public opinion. As one commenter on the ‘AI Explained’ channel noted, “We are mindlessly hurtling towards a world of noise where nothing can be trusted or makes any sense.” This dystopian vision highlights the urgent need for regulatory frameworks and technological solutions to address the risks posed by AI-generated content.

Some experts are calling for stricter regulations and the development of new technologies to help detect and counteract deepfakes. Demis Hassabis, CEO of Google DeepMind, recently emphasized the importance of proactive measures: “We need to be proactive in addressing these issues. The technology is advancing quickly, and if we’re not careful, it could outpace our ability to control it.”

A Turning Point or Just Another Step?

The debate over Grok-2’s significance is far from settled. Some view it as a harbinger of a new era of AI-driven innovation, while others see it as just another model in an increasingly crowded field. As one skeptic on the ‘AI Explained’ channel remarked, “How can we really judge the importance of Grok-2 when there’s no transparency about how it works or what it’s truly capable of? Without that, it’s just another black box.”

Despite these reservations, Grok-2’s release is undeniably a moment of interest in the AI landscape. The model’s capabilities, as demonstrated through early benchmark performances, suggest it could play a significant role in shaping the future of AI. However, this potential is tempered by the ongoing challenges in AI development, particularly around ethics, transparency, and the limits of scaling.

The ethical implications of models like Grok-2 cannot be overstated. As AI continues to advance, the line between reality and digital fabrication becomes increasingly blurred, raising concerns about trust and authenticity in the digital age. The potential for real-time deepfakes, coupled with the model’s capabilities, presents both opportunities and risks that society must grapple with sooner rather than later.

Ultimately, Grok-2’s legacy will depend on how these challenges are addressed. Will the AI community find ways to harness the power of large language models while ensuring they are used responsibly? Or will Grok-2 and its successors become symbols of an era where technological advancement outpaced our ability to manage its consequences?

As we stand at this crossroads, the future of AI remains uncertain. Grok-2 might just be one of many signposts along the way, pointing to the immense possibilities—and dangers—of what lies ahead.

]]>
606887
Zed Editor Adds Anthropic-Powered AI Features https://www.webpronews.com/zed-editor-adds-anthropic-powered-ai-features/ Wed, 21 Aug 2024 19:30:13 +0000 https://www.webpronews.com/?p=606716 Zed, the text editor taking the development world by storm, has announced new AI features powered by Anthropic’s Claude.

Zed is a new text editor written entirely in Rust, benefiting from the speed, security, and other features the language provides. Zed has been gaining in popularity, with a Linux version of the text editor recently being released.

The company is now working with Anthropic to bring AI-powered features to the text editor, according to a blog post. Nathan Sobo, Zed founder, said the company has been looking for ways to integrate LLMs in a way that enhanced productivity.

In the two years since LLMs came onto our radar, we’ve been focused on building out the core of Zed: a fast, reliable text editor with the features developers need. Meanwhile, we’ve been quietly experimenting with integrating LLMs into our own workflows. Not as a flashy gimmick, but as a practical tool to enhance our productivity working on a complex, real-world codebase.

It appears Anthropic is a Zed fan, approaching the company to discuss a integration.

As we refined our AI integration, we caught the attention of some unexpected allies. Engineers at Anthropic, one of the world’s leading AI companies, discovered Zed and quickly saw the value of our raw, text-centric interface that puts minimal separation between the user and the language model. Their enthusiasm was validating, and our conversations sparked a dialogue that quickly evolved into a collaboration.

Now, we’re ready to introduce Zed AI, a hosted service providing convenient and performant support for AI-enabled coding in Zed, powered by Anthropic’s Claude 3.5 Sonnet and accessible just by signing in. We also worked with Anthropic to optimize Zed for implement their new Prompt Caching beta, leading to lightning-fast responses even with thousands of lines of code included in the context window while reducing cost.

Zed AI is composed of two components, one of which is the assistant panel.

The assistant panel is where you interact with AI models in Zed, but it’s not your typical chat interface. It’s a full-fledged text editor that exposes the entire LLM request. Code snippets, conversation history, file contents—it’s all there, and it’s all just text. You can observe, edit, and refine any part of the request using familiar coding tools, giving you full transparency and control over every interaction.

The second component is inline transformations.

Inline transformations, activated with ctrl-enter, allow you to transform and generate code via natural language prompts. What sets them apart is their precision and responsiveness.

To give you fast feedback, we’ve implemented a custom streaming diff protocol that works with Zed’s CRDT-based buffers to deliver edits as soon as they’re streamed from the model. You see the model’s output token by token, allowing you to read and react to changes as they happen. This low-latency streaming creates a fluid, interactive coding experience that keeps you engaged and in control throughout the process.

Inline transformations in Zed use the context you’ve built in the assistant panel. There’s no hidden system prompt-you see and control every input shaping the model’s output. This transparency lets you fine-tune the model’s behavior and improve your skills in AI-assisted coding.

Sobo says the company is working on additional features for Zed AI, including workflows for complex transformations and tools to efficiently build context. Sobo invites developers to help craft the future of Zed AI.

Zed AI embodies our belief in open, collaborative software development. We’ve created a transparent, extensible environment that empowers you to harness AI on your own terms, keeping you firmly in control of your tools and workflows.

We invite you to try Zed AI and become part of this journey. Experiment with custom slash commands, fine-tune prompts, and push boundaries. Share your innovations as extensions or as contributions to the Zed repository.

With Zed AI, you’re in the driver’s seat, directing AI’s potential within the familiar realm of text. Together, we’ll build an AI-assisted development experience that amplifies your creativity and adapts to your unique coding style. We’re excited to see what our community will create.

Anthropic is also helping to further Zed development, with the AI firm’s Rust engineers are actively contributing to Zed’s open-source codebase.

Those interested in trying Zed can download versions for macOS and Linux here.

]]>
606716
California Partners with Nvidia to Revolutionize AI Training in Community Colleges https://www.webpronews.com/california-partners-with-nvidia-to-revolutionize-ai-training-in-community-colleges/ Fri, 09 Aug 2024 20:23:11 +0000 https://www.webpronews.com/?p=606359 In a groundbreaking move to fortify its position at the forefront of technological innovation, California has partnered with Nvidia to bring cutting-edge artificial intelligence (AI) resources to the state’s expansive community college system. The partnership, formalized by Governor Gavin Newsom and Nvidia CEO Jensen Huang, represents a significant stride in equipping students, educators, and workers with the skills necessary to thrive in an increasingly AI-driven world.

A Strategic Alliance for AI Education

The collaboration is set to transform the educational landscape across California’s 116 community colleges, which serve over two million students. Under the terms of the agreement, Nvidia will provide access to its state-of-the-art AI tools, including hardware, software, and specialized training materials. These resources will be integrated into college curriculums, focusing on the practical applications of AI in high-demand sectors such as technology, healthcare, and finance.

“California’s world-leading companies are pioneering AI breakthroughs, and it’s essential that we create more opportunities for Californians to get the skills to utilize this technology and advance their careers,” Governor Newsom said during the signing ceremony. This initiative aligns with the state’s broader goals of fostering innovation and ensuring that all Californians can benefit from advancements in AI.

Empowering the Workforce of Tomorrow

The partnership focuses on equipping the next generation of workers with the tools they need to succeed in a rapidly changing job market. Nvidia will offer AI-focused certifications, workshops, and boot camps to help students and faculty stay ahead of industry trends. Additionally, the company will support the development of AI laboratories across community colleges, enabling hands-on learning experiences that will prepare students for the future workforce.

“We’re in the early stages of a new industrial revolution that will transform trillion-dollar industries around the world,” Nvidia’s Jensen Huang said. “Together with California, Nvidia will train 100,000 students, college faculty, developers, and data scientists to harness this technology to prepare California for tomorrow’s challenges and unlock prosperity throughout the state.”

Addressing Equity and Inclusion

One of the key aspects of this initiative is its focus on equitable access to AI education. The collaboration aims to bridge the gap for underserved populations by ensuring that students from all backgrounds have the opportunity to gain industry-aligned AI skills. Sonya Christian, Chancellor of California Community Colleges, emphasized this commitment: “Our approach prioritizes equitable access to AI teaching and learning enhancements that will lift up underserved populations.”

This emphasis on inclusivity reflects California’s broader commitment to using technology for social and economic advancement. The partnership hopes to create a more inclusive workforce prepared to tackle future challenges by providing AI education and resources to a diverse student body.

A Vision for the Future

The California-Nvidia partnership is part of a larger vision to position the state as a global leader in AI innovation. The initiative builds on Governor Newsom’s 2023 executive order, which called for the responsible use of AI to benefit all Californians. This collaboration not only sets a new standard for public-private partnerships but also highlights the critical role that education will play in shaping the future of AI.

As AI continues to evolve, the importance of equipping the workforce with the necessary skills cannot be overstated. The California-Nvidia partnership is a bold step toward ensuring that the state remains at the cutting edge of technological advancement while also promoting equity and opportunity for all its residents.

With this initiative, California is preparing for the future and actively shaping it.

]]>
606359
FTC’s Lina Khan Sees Open AI Models As The Answer To AI Monopolies https://www.webpronews.com/ftcs-lina-khan-sees-open-ai-models-as-the-answer-to-ai-monopolies/ Fri, 26 Jul 2024 22:21:01 +0000 https://www.webpronews.com/?p=606015 Federal Trade Commission Chair Lina Khan has vocalized her support for open AI models, saying they could prove the key to preventing AI monopolies.

According to Bloomberg, Khan made the comments at Y Combinator in San Francisco.

“There’s tremendous potential for open-weight models to promote competition,” Khan said. “Open-weight models can liberate startups from the arbitrary whims of closed developers and cloud gatekeepers.”

Khan’s comments come at a time when regulators on both sides of the Atlantic are growing increasingly wary of Big Tech. AI companies have done little to stave off such concerns, with accusations they plagiarize content, throttle organizations’ servers as they scrape them, and show little regard for the potential danger AI may pose.

In view those issues, many lawmakers are concerned about a future where AI development and breakthroughs are largely controlled by a handful of companies.

One notable exception in the industry is Meta’s Llama AI model, which the company has made available as open-source software. The company explained its reasons in a blog post announcing Llama 3:

We’re committed to the continued growth and development of an open AI ecosystem for releasing our models responsibly. We have long believed that openness leads to better, safer products, faster innovation, and a healthier overall market. This is good for Meta, and it is good for society. We’re taking a community-first approach with Llama 3, and starting today, these models are available on the leading cloud, hosting, and hardware platforms with many more to come.

With Khan’s comments, LLama and other open models may see an uptick in use.

]]>
606015
Apple Signs Biden Administration’s AI Safety Guidelines https://www.webpronews.com/apple-signs-biden-administrations-ai-safety-guidelines/ Fri, 26 Jul 2024 19:51:18 +0000 https://www.webpronews.com/?p=606010 In preparation for the release of its Apple Intelligence, the iPhone make has voluntarily signed the Biden Administration’s AI safety guidelines.

The White House announced the news in a press release:

Nine months ago, President Biden issued a landmark Executive Order to ensure that America leads the way in seizing the promise and managing the risks of artificial intelligence (AI).

This Executive Order built on the voluntary commitments he and Vice President Harris received from 15 leading U.S. AI companies last year. Today, the administration announced that Apple has signed onto the voluntary commitments, further cementing these commitments as cornerstones of responsible AI innovation.

Apple is widely considered to be a significant factor in the AI industry, thanks largely to its penchant for making high-tech solutions approach to the average user, as well as the huge user base that it can leverage.

With the announcement of Apple Intelligence, many critics and experts say Apple has done more to make the case for AI’s usefulness to the average user than most other companies combined. In view of the role Apple will likely play, it’s good to see the company’s continued commitment to safe AI development and deployment.

]]>
606010
xAI Goes Its Own Way Instead of Depending On Oracle https://www.webpronews.com/xai-goes-its-own-way-instead-of-depending-on-oracle/ Wed, 10 Jul 2024 23:00:00 +0000 https://www.webpronews.com/?p=605601 Elon Musk announced that his AI startup, xAI, will deploy Nvidia H100 systems on its own rather than continuing to use Oracle.

Musk’s xAI originally tapped Oracle to help it deploy 24,000 H100s that were used to train its Grok 2 model. According to Musk, however, the company plans to go its own way, building out its own cluster containing some 100,000 H100s. Musk framed the decision in the context of needing to leapfrog its AI rivals, with controlling its own cluster being the key to doing so.

xAI contracted for 24k H100s from Oracle and Grok 2 trained on those. Grok 2 is going through finetuning and bug fixes. Probably ready to release next month.

xAI is building the 100k H100 system itself for fastest time to completion. Aiming to begin training later this month. It will be the most powerful training cluster in the world by a large margin.

The reason we decided to do the 100k H100 and next major system internally was that our fundamental competitiveness depends on being faster than any other AI company. This is the only way to catch up.

Oracle is a great company and there is another company that shows promise also involved in that OpenAI GB200 cluster, but, when our fate depends on being the fastest by far, we must have our own hands on the steering wheel, rather than be a backseat driver.

Elon Musk (@elonmusk) | July 9, 2024

The move is a blow to Oracle. As Investors.com points out, Oracle founder Larry Ellison touted its relationship with xAI in a recent quarterly earnings call, saying his company was working to secure more H100s for the startup.

“We gave them quite a few,” Ellison said at the time. “But they wanted more, and we are in the process of getting them more.”

]]>
605601
Anthropic Adds the Ability to Evaluate Prompts https://www.webpronews.com/anthropic-adds-the-ability-to-evaluate-prompts/ Wed, 10 Jul 2024 12:03:00 +0000 https://www.webpronews.com/?p=605606 Anthropic is making it easier for developers to generate high-quality prompts, adding prompt evaluation to the Anthropic Console.

Prompts are an important part of the AI development process, and can have a major impact on the results, as Anthropic says in a blog post announcing the new feature:

When building AI-powered applications, prompt quality significantly impacts results. But crafting high quality prompts is challenging, requiring deep knowledge of your application’s needs and expertise with large language models. To speed up development and improve outcomes, we’ve streamlined this process to make it easier for users to produce high quality prompts.

You can now generate, test, and evaluate your prompts in the Anthropic Console. We’ve added new features, including the ability to generate automatic test cases and compare outputs, that allow you to leverage Claude to generate the very best responses for your needs.

Anthropic says users can generate prompts simply by describing a task to Claude. Using the Claude 3.5 Sonnet engine, Claude will use the description its given to generate a high-quality prompt.

The new Evaluate feature makes it much easier to test prompts against real-world inputs.

Testing prompts against a range of real-world inputs can help you build confidence in the quality of your prompt before deploying it to production. With the new Evaluate feature you can do this directly in our Console instead of manually managing tests across spreadsheets or code.

Manually add or import new test cases from a CSV, or ask Claude to auto-generate test cases for you with the ‘Generate Test Case’ feature. Modify your test cases as needed, then run all of the test cases in one click. View and adjust Claude’s understanding of the generation requirements for each variable to get finer-grained control over the test cases Claude generates.

Anthropic is already the leading OpenAI competitor, with its Claude 3.5 besting OpenAI’s GPT-4o in a range of tests. With the new features aimed at improving the quality of prompts, Anthropic continues to push AI development forward.

]]>
605606
Microsoft Azure AI Skirts OpenAI’s China Ban https://www.webpronews.com/microsoft-azure-ai-skirts-openais-china-ban/ Mon, 08 Jul 2024 16:03:21 +0000 https://www.webpronews.com/?p=605564 Microsoft Azure has managed to avoid OpenAI’s ban on providing API access in China, a major win for the Redmond company’s cloud and AI efforts.

OpenAI announced in late June that it would block API traffic from countries that were not on its “supported countries and territories” list. Users on the company’s forums reported receiving emails from the company informing them of the policy.

China was conspicuously absent from the long list of supported countries, meaning that Chinese developers will not have access to the company’s API for development. According to The Information, however, there is a significant workaround that runs straight through Microsoft Azure AI.

Developers in China who want to take advantage of OpenAI’s models can still do so if they sign up for an Azure account. Because of Microsoft and OpenAI’s close relationship, this gives developers access to the AI firm’s AI models through Microsoft’s services.

According to the outlet, the exception works because Azure China is a joint venture with Chinese company 21Vianet. Multiple customers confirmed to The Information that they had full access to OpenAI models within Azure.

Given the importance of the Chinese market, the revelation is good news for Microsoft, OpenAI, and Chinese AI developers.

]]>
605564
Microsoft Details ‘Skeleton Key’ AI Model Jailbreak https://www.webpronews.com/microsoft-details-skeleton-key-ai-model-jailbreak/ Wed, 03 Jul 2024 17:55:54 +0000 https://www.webpronews.com/?p=605529 Microsoft is detailing a jailbreak, dubbed “Skeleton Key,” that can be used to trick an AI model into operation outside of its parameters.

AI models are designed to operate within strictly defined parameters that ensure the responses it gives are not offensive and do not cause harm. This is something AI firms have struggled with, with AI models sometimes going beyond their parameters and stirring up controversy in the process.

According to Microsoft Security, there is a newly discovered jailbreak attack—Skeleton Key—that impacts multiple AI models from various firms (hence the name).

This AI jailbreak technique works by using a multi-turn (or multiple step) strategy to cause a model to ignore its guardrails. Once guardrails are ignored, a model will not be able to determine malicious or unsanctioned requests from any other. Because of its full bypass abilities, we have named this jailbreak technique Skeleton Key.

This threat is in the jailbreak category, and therefore relies on the attacker already having legitimate access to the AI model. In bypassing safeguards, Skeleton Key allows the user to cause the model to produce ordinarily forbidden behaviors, which could range from production of harmful content to overriding its usual decision-making rules. Like all jailbreaks, the impact can be understood as narrowing the gap between what the model is capable of doing (given the user credentials, etc.) and what it is willing to do. As this is an attack on the model itself, it does not impute other risks on the AI system, such as permitting access to another user’s data, taking control of the system, or exfiltrating data.

Microsoft says it has already made a number of updates to its Copilot AI assistants and other LLM technology in an effort to mitigate the attack. The company says customers should consider the following actions to implement their own AI system design:

  • Input filtering: Azure AI Content Safety detects and blocks inputs that contain harmful or malicious intent leading to a jailbreak attack that could circumvent safeguards.
  • System message: Prompt engineering the system prompts to clearly instruct the large language model (LLM) on appropriate behavior and to provide additional safeguards. For instance, specify that any attempts to undermine the safety guardrail instructions should be prevented (read our guidance on building a system message framework here).
  • Output filtering: Azure AI Content Safety post-processing filter that identifies and prevents output generated by the model that breaches safety criteria.
  • Abuse monitoring: Deploying an AI-driven detection system trained on adversarial examples, and using content classification, abuse pattern capture, and other methods to detect and mitigate instances of recurring content and/or behaviors that suggest use of the service in a manner that may violate guardrails. As a separate AI system, it avoids being influenced by malicious instructions. Microsoft Azure OpenAI Service abuse monitoring is an example of this approach.

The company says its Azure AI tools already help customers protect against this type of attack as well:

Microsoft provides tools for customers developing their own applications on Azure. Azure AI Content Safety Prompt Shields are enabled by default for models hosted in the Azure AI model catalog as a service, and they are parameterized by a severity threshold. We recommend setting the most restrictive threshold to ensure the best protection against safety violations. These input and output filters act as a general defense not only against this particular jailbreak technique, but also a broad set of emerging techniques that attempt to generate harmful content. Azure also provides built-in tooling for model selection, prompt engineering, evaluation, and monitoring. For example, risk and safety evaluations in Azure AI Studio can assess a model and/or application for susceptibility to jailbreak attacks using synthetic adversarial datasets, while Microsoft Defender for Cloud can alert security operations teams to jailbreaks and other active threats.

With the integration of Azure AI and Microsoft Security (Microsoft Purview and Microsoft Defender for Cloud) security teams can also discover, protect, and govern these attacks. The new native integration of Microsoft Defender for Cloud with Azure OpenAI Service, enables contextual and actionable security alerts, driven by Azure AI Content Safety Prompt Shields and Microsoft Defender Threat Intelligence. Threat protection for AI workloads allows security teams to monitor their Azure OpenAI powered applications in runtime for malicious activity associated with direct and in-direct prompt injection attacks, sensitive data leaks and data poisoning, or denial of service attacks.

The Skeleton Key attack underscores the ongoing challenges facing companies as AI becomes more widely used. While it can be a valuable tool for cybersecurity, it can also open up entirely new attack vectors.

]]>
605529
Anthropic Calls For A New Way To Evaluate AI https://www.webpronews.com/anthropic-calls-for-a-new-way-to-evaluate-ai/ Wed, 03 Jul 2024 13:00:00 +0000 https://www.webpronews.com/?p=605519 Anthropic, one of the leaders in AI development, is calling for proposals to help “fund evaluations developed by third-party organizations.”

Properly evaluating AI’s potential is a growing challenge for AI firms as the technology evolves. Not only is it challenging to properly evaluate an AI’s capabilities, but there are also concerns with evaluating the various safety issues involved.

Anthropic has increasingly been setting itself apart in the AI field, not only for its powerful Claude model that is currently beating OpenAI’s GPT-4o, but also for its safety-first approach to AI. In fact, the company was founded by OpenAI executives that were concerned with the direction of OpenAI, and the company has continued to attract disillusioned OpenAI engineers. The most notable recent example is Jan Leike, who left OpenAI after the safety team he co-lead was disbanded.

With that background, it’s not surprising that Anthropic is interested in developing and discovering new and better ways to properly evaluate AI. The company outlines its highest priority areas of focus:

  • AI Safety Level assessments
  • Advanced capability and safety metrics
  • Infrastructure, tools, and methods for developing evaluations

The company outlines a number of AI Safety Levels (ASLs) that is concerned with, including cybersecurity; chemical, biological, radiological, and nuclear (CBRN) risks; model autonomy; national security risks; and misalignment risks. In all of these areas, the company is concerned with the risk that AI could be used to aid individuals in doing harm.

We’re particularly interested in capabilities that, if automated and scaled, could pose significant risks to critical infrastructure and economically valuable systems at levels approaching advanced persistent threat actors.

We’re prioritizing evaluations that assess two critical capabilities: a) the potential for models to significantly enhance the abilities of non-experts or experts in creating CBRN threats, and b) the capacity to design novel, more harmful CBRN threats.

AI systems have the potential to significantly impact national security, defense, and intelligence operations of both state and non-state actors. We’re committed to developing an early warning system to identify and assess these complex emerging risks.

Anthropic reveals a fascinating, and terrifying, observation about current AI models, what the company identifies as “misalignment risks.”

Our research shows that, under some circumstances, AI models can learn dangerous goals and motivations, retain them even after safety training, and deceive human users about actions taken in their pursuit.

The company says this represents a major danger moving forward as AI models become more advanced.

These abilities, in combination with the human-level persuasiveness and cyber capabilities of current AI models, increases our concern about the potential actions of future, more-capable models. For example, future models might be able to pursue sophisticated and hard-to-detect deception that bypasses or sabotages the security of an organization, either by causing humans to take actions they would not otherwise take or exfiltrating sensitive information.

Anthropic goes on to highlight its desire to improve evaluation methods to address bias issues, something that has been a significant challenge in training existing AI models.

Evaluations that provide sophisticated, nuanced assessments that go beyond surface-level metrics to create rigorous assessments targeting concepts like harmful biases, discrimination, over-reliance, dependence, attachment, psychological influence, economic impacts, homogenization, and other broad societal impacts.

The company also wants to ensure AI benchmarks support multiple languages, something that is not currently the case. New evaluation methods should also be able to “detect potentially harmful model outputs,” such as “attempts to automate cyber incidents.” The company also wants the new evaluation methods to better determine AI’s ability to learn, especially in the sciences.

Anthropic’s Criteria

Parties interested in submitting a proposal should keep the company’s 10 requirements in mind:

  1. Sufficiently difficult: Evaluations should be relevant for measuring the capabilities listed for levels ASL-3 or ASL-4 in our Responsible Scaling Policy, and/or human-expert level behavior.
  2. Not in the training data: Too often, evaluations end up measuring model memorization because the data is in its training set. Where possible and useful, make sure the model hasn’t seen the evaluation. This helps indicate that the evaluation is capturing behavior that generalizes beyond the training data.
  3. Efficient, scalable, ready-to-use: Evaluations should be optimized for efficient execution, leveraging automation where possible. They should be easily deployable using existing infrastructure with minimal setup.
  4. High volume where possible: All else equal, evaluations with 1,000 or 10,000 tasks or questions are preferable to those with 100. However, high-quality, low-volume evaluations are also valuable.
  5. Domain expertise: If the evaluation is about expert performance on a particular subject matter (e.g. science), make sure to use subject matter experts to develop or review the evaluation.
  6. Diversity of formats: Consider using formats that go beyond multiple choice, such as task-based evaluations (for example, seeing if code passes a test or a flag is captured in a CTF), model-graded evaluations, or human trials.
  7. Expert baselines for comparison: It is often useful to compare the model’s performance to the performance of human experts on that domain.
  8. Good documentation and reproducibility: We recommend documenting exactly how the evaluation was developed and any limitations or pitfalls it is likely to have. Use standards like the Inspect or the METR standard where possible.
  9. Start small, iterate, and scale: Start by writing just one to five questions or tasks, run a model on the evaluation, and read the model transcripts. Frequently, you’ll realize the evaluation doesn’t capture what you want to test, or it’s too easy.
  10. Realistic, safety-relevant threat modeling: Safety evaluations should ideally have the property that if a model scored highly, experts would believe that a major incident could be caused. Most of the time, when models have performed highly, experts have realized that high performance on that version of the evaluation is not sufficient to worry them.

Those interested in submitting a proposal, and possibly working long-term with Anthropic, should use this application form.

OpenAI has been criticized for for a lack of transparency that has led many to believe the company has lost its way and is no longer focused on its one-time goal of safe AI development. Anthropic’s willingness to engage the community and industry is a refreshing change of pace.

]]>
605519
Meta Changes Its Approach To AI Labels On Photos After Backlash https://www.webpronews.com/meta-changes-its-approach-to-ai-labels-on-photos-after-backlash/ Tue, 02 Jul 2024 15:10:58 +0000 https://www.webpronews.com/?p=605506 Meta announced it is changing its approach to its “Made with AI” labels on photos after it incorrectly identified photos taken by photographers as AI-generated.

Labeling AI content has become a growing concern for online platforms, as well as regulators, as AI-generated content has become so realistic that it could easily be used to create false narratives. Meta announced in April plans to label AI content with a “Made with AI” label. Unfortunately, it’s algorithm for identifying AI content had some issues, with photos taken by human photographers being improperly labeled.

The company says it has made changes to address the issue.

We want people to know when they see posts that have been made with AI. Earlier this year, we announced a new approach for labeling AI-generated content. An important part of this approach relies on industry standard indicators that other companies include in content created using their tools, which help us assess whether something is created using AI.

Like others across the industry, we’ve found that our labels based on these indicators weren’t always aligned with people’s expectations and didn’t always provide enough context. For example, some content that included minor modifications using AI, such as retouching tools, included industry standard indicators that were then labeled “Made with AI.” While we work with companies across the industry to improve the process so our labeling approach better matches our intent, we’re updating the “Made with AI” label to “AI info” across our apps, which people can click for more information.

According to CNET, photographer Pete Souza said cropping tools appear to be one of the culprits. Because such tools add information to images, it seems that Meta’s algorithm was incorrectly identifying that added information and taking it as an indication the images were AI-generated.

The entire issue demonstrates the growing challenges associated with correctly identifying AI-generated content. For years, experts have warned about the potential havoc deepfakes could cause, impacting everything from people’s personal lives to business to politics.

Interestingly, OpenAI shuttered its own AI-content detection tool in early 2024, saying at the time that such tools don’t work:

While some (including OpenAI) have released tools that purport to detect AI-generated content, none of these have proven to reliably distinguish between AI-generated and human-generated content.

It remains to be seen if Meta will be able to reliably identify AI-generated images, or if it will suffer the same issues that led OpenAI to throw in the towel.

]]>
605506
Microsoft AI CEO Says Online Content Is Fair Game https://www.webpronews.com/microsoft-ai-ceo-says-online-content-is-fair-game/ Sat, 29 Jun 2024 17:57:37 +0000 https://www.webpronews.com/?p=605457 Microsoft AI CEO Mustafa Suleyman has raised eyebrows, saying online content is fair game for training AI models.

Content ownership has become one of the most controversial and contentious aspects of AI development. AI models need vast amounts of data for training and most AI firms have used online data as their source, some establishing paid content deals, and others scraping data without paying for it.

According to Suleyman, content found online should be governed by fair use.

“I think that with respect to content that is already on the open web, the social contract of that content since the 1990s has been it is fair use,” he said in an interview with CNBC, via The Register. “Anyone can copy it, recreate with it, reproduce with it. That has been freeware, if you like. That’s been the understanding.”

The one exception Suleyman made is for websites and publishers that explicitly forbid content scraping.

“There’s a separate category where a website or publisher or news organization had explicitly said, ‘do not scrape or crawl me for any other reason than indexing me,’ so that other people can find that content,” he explained. “But that’s the gray area. And I think that’s going to work its way through the courts.”

Unfortunately, not all AI firms respect ‘do not crawl’ requests. AWS is investigating Perplexity AI over accusations it is scraping websites in violation of their terms, using AWS’ cloud platform to do so.

As Suleyman points out, the legality of the practice will ultimately be decided in the courts but, in the meantime, content ownership will continue to be a hotly debated topic.

]]>
605457
AWS Investigates Perplexity AI Scraping Allegations https://www.webpronews.com/aws-investigates-perplexity-ai-scraping-allegations/ Fri, 28 Jun 2024 15:42:05 +0000 https://www.webpronews.com/?p=605451 AWS is investigating allegations that Perplexity AI used the company’s cloud platform for unauthorized website scraping.

Content ownership is one of the biggest legal and ethical challenges facing AI firms. Some AI firms have committed to honoring the Robots Exclusion Protocol, a standard that defines what web pages search engines should index or scrape, and which should be ignored.

Perplexity AI has been accused of ignoring the protocol and scraping sites without permission, according to Wired. As a result, sources within AWS confirmed to the outlet that it was investigating the AI firm, since AWS requires that its cloud customers adhere to the Robots Exclusion Protocol.

“AWS’s terms of service prohibit customers from using our services for any illegal activity, and our customers are responsible for complying with our terms and all applicable laws,” the spokesperson said in a statement to Wired.

Perplexity has already faced accusations of stealing articles and plagiarism. If the AWS investigation goes against the AI firm, it remains to be seen what action the cloud provider may take.

]]>
605451
Reddit Is Updating Its Policies to Crack Down On AI Scraping https://www.webpronews.com/reddit-is-updating-its-policies-to-crack-down-on-ai-scraping/ Wed, 26 Jun 2024 19:28:25 +0000 https://www.webpronews.com/?p=605429 Reddit is updating its policies in an apparent effort to crack down on AI companies scraping the site for content to train AI models.

Reddit is a popular place for AI companies to scrape, thanks to the large quantity of user-generated content on a vast array of subjects. Reddit has signed a deal with Google allowing the company to use the site’s content, but other companies appear to be continuing their efforts to scrape the site.

The company says it will make changes to address the issue.

In the coming weeks, we’ll update our Robots Exclusion Protocol (robots.txt file), which gives high-level instructions about how we do and don’t allow Reddit to be crawled by third parties. Along with our updated robots.txt file, we will continue rate-limiting and/or blocking unknown bots and crawlers from accessing reddit.com. This update shouldn’t impact the vast majority of folks who use and enjoy Reddit. Good faith actors – like researchers and organizations such as the Internet Archive – will continue to have access to Reddit content for non-commercial use.

Mark Graham, Director, Wayback Machine at Internet Archive, praised Reddit’s position.

“The Internet Archive is grateful that Reddit appreciates the importance of helping to ensure the digital records of our times are archived and preserved for future generations to enjoy and learn from,” said Graham. “Working in collaboration with Reddit we will continue to record and make available archives of Reddit, along with the hundreds of millions of URLs from other sites we archive every day.”

Reddit emphasized that organizations must abide by its policies.

Anyone accessing Reddit content must abide by our policies, including those in place to protect redditors. We are selective about who we work with and trust with large-scale access to Reddit content. Organizations looking to access Reddit content can head over to our guide to accessing Reddit Data.

]]>
605429
LibreChat Offers ‘Every AI’ While Protecting Your Data and Privacy https://www.webpronews.com/librechat-offers-every-ai-while-protecting-your-data-and-privacy/ Tue, 25 Jun 2024 00:51:00 +0000 https://www.webpronews.com/?p=605374 LibreChat announced it now offers “every AI in one place, built for everyone,” while giving users more control over their security and privacy.

AI models are growing in popularity, but not all users are comfortable giving the companies behind them access to their data. LibreChat is an open-source AI platform that gives user access to multiple models in a centralized hub.

LibreChat is a free, open-source AI chat platform that empowers you to harness the capabilities of cutting-edge language models from multiple providers in a unified interface. With its vast customization options, innovative enhancements, and seamless integration of AI services, LibreChat offers an unparalleled conversational experience.

LibreChat is an enhanced, open-source ChatGPT clone that brings together the latest advancements in AI technology. It serves as a centralized hub for all your AI conversations, providing a familiar, user-friendly interface enriched with advanced features and customization capabilities.

The project emphasizes several design principles, not the least of which are privacy and security.

  1. User-Friendly Interface: Inspired by the familiar ChatGPT UI, LibreChat offers a clean and intuitive layout, making it easy for users to engage with AI assistants.
  2. Multimodal Conversations: LibreChat supports multimodal conversations, allowing you to upload and analyze images, chat with files, and leverage advanced agent capabilities powered by AI models like GPT-4 Claude and Gemini Vision.
  3. Extensibility: With its plugin architecture and open-source nature, LibreChat encourages the development of custom extensions and integrations, enabling users to tailor the platform to their specific needs.
  4. Privacy and Security: LibreChat prioritizes privacy and security by offering secure authentication, moderation tools, and the ability to self-host the application.

LibreChat is an intriguing entrance in the AI market, one that demonstrates the ingenuity of the open-source community.

]]>
605374
The Future of Developer Recruitment: AI, Automation, and Beyond https://www.webpronews.com/developer-recruitment/ Tue, 04 Jun 2024 16:10:08 +0000 https://www.webpronews.com/?p=605025 The landscape of developer recruitment is undergoing a transformative shift, propelled by rapid technological advancements and changing industry dynamics. As businesses increasingly rely on digital solutions, the demand for skilled developers has skyrocketed, making the recruitment process more competitive and complex. In this evolving scenario, Artificial Intelligence (AI), automation, and other technological innovations are emerging as key players, offering new ways to streamline and enhance the recruitment process. These technologies not only promise to improve efficiency but also aim to redefine how companies attract, evaluate, and hire tech talent. 

This evolution in recruitment strategies invites companies to leverage cutting-edge tools on a tech recruitment platform, such as a Huntly, and stay ahead in the talent acquisition game.

Key Takeaway

  • AI is the New Headhunter: AI’s taking over the boring part of looking through resumes and picking out the best matches based on what the job needs. It’s like having a super-smart assistant who doesn’t get tired or play favorites, making sure everyone gets a fair shot.
  • Interviews Go Digital: Forget about traveling for interviews. Now, you can meet candidates and test their skills online through virtual reality challenges or video calls. It opens up the whole world as your talent pool, letting you find the perfect fit no matter where they live.
  • Learning Never Stops: The smartest companies aren’t just looking for what you know now; they’re interested in how fast you can learn new stuff. With tech changing all the time, having platforms that help everyone keep up and get better is becoming a big part of picking the right people.

AI-Powered Candidate Screening

AI is revolutionizing the initial stages of recruitment by automating the screening of candidates. Machine learning algorithms can sift through hundreds of applications, identifying those that best match the job requirements based on skills, experience, and potential. This not only saves valuable time but also minimizes human bias, ensuring a more diverse and qualified candidate pool progresses to the interview stage.

Enhanced Candidate Matching

Beyond screening, AI systems are becoming increasingly sophisticated at matching candidates with job vacancies. By analyzing data points across previous successful hires and ongoing performance metrics, these systems can predict candidate success more accurately. This approach not only improves the quality of hires but also contributes to longer-term employee satisfaction and retention.

Automated Scheduling and Communication

Automation is streamlining administrative tasks such as interview scheduling and candidate communication. Chatbots and AI-driven platforms can handle queries, provide updates, and manage scheduling without human intervention, enhancing the candidate experience by ensuring prompt and personalized interaction throughout the recruitment process.

Virtual Interviews and Assessments

The rise of virtual interviews and assessment tools is making geographical boundaries irrelevant. Online coding tests, virtual reality (VR) environments for real-world problem-solving, and video interviews allow for a comprehensive evaluation of candidates’ technical and soft skills without the need for physical presence. This opens up a global talent pool, enabling companies to attract and assess candidates from across the world.

Predictive Analytics

Predictive analytics is shaping the future of recruitment by forecasting hiring needs and candidate success. By analyzing trends, skills evolution, and company growth patterns, businesses can anticipate future recruitment needs and build strategic talent pipelines. This proactive approach ensures companies are always prepared to meet their developmental and technological challenges with the right talent.

Continuous Learning and Development Platforms

Recognizing the fast-paced nature of technological advancement, recruitment strategies are increasingly focusing on candidates’ potential for continuous learning. Platforms that offer ongoing skill assessment and development opportunities are becoming integral to the recruitment process, allowing companies to not only hire for current needs but also invest in the future growth of their employees.

Ethical Considerations and Human Oversight

As technology reshapes recruitment, ethical considerations, and human oversight remain paramount. Ensuring AI and automation are used responsibly to avoid biases and protect candidate privacy is crucial. Companies must balance technological efficiency with a human touch, ensuring that recruitment processes remain fair, transparent, and respectful of candidates’ rights.

The future of developer recruitment is intricately linked to AI, automation, and technological innovations. These advancements promise a more efficient, accurate, and global recruitment process, allowing companies to meet the growing demand for tech talent effectively. 

However, navigating this future requires a mindful approach to leveraging technology while maintaining ethical standards and human connections. As the industry continues to evolve, embracing these innovations while staying grounded in human-centric recruitment practices will be key to attracting and retaining the best tech talent.

]]>
605025
IBM’s Vision for the Future of AI: Open and Collaborative https://www.webpronews.com/ibms-vision-for-the-future-of-ai-open-and-collaborative/ Thu, 23 May 2024 19:49:29 +0000 https://www.webpronews.com/?p=604889 SAN FRANCISCO – In a keynote address at IBM’s Think 2024 conference, IBM SVP and Director of Research Darío Gil outlined groundbreaking advancements in artificial intelligence (AI) that promise to transform how enterprises leverage technology. The event, held at the bustling Moscone Center, gathered industry leaders, tech enthusiasts, and innovators eager to hear about the future of AI from one of its most prominent voices.

The future of AI is open,” declared Gil, emphasizing the importance of open-source innovation and collaborative efforts in the AI landscape. He urged businesses to adopt open strategies to maximize the potential of their AI systems, arguing that such an approach not only fosters innovation but also ensures flexibility and adaptability.

Embracing Open Source

Open is about innovating together, not in isolation,” Gil said. By choosing open-source frameworks, companies can decide which models to use, what data to integrate, and how to adapt AI to their specific needs. Gil argued that this collaborative approach is essential for the evolution of AI to meet the diverse aspirations of various industries.

The strength of open source lies in its ability to foster a community-driven ecosystem where innovation can thrive unencumbered by proprietary constraints. Gil pointed to the success of IBM’s own Granite family of models, designed to handle tasks ranging from coding to time series analysis and geospatial data processing. These models, released under an Apache 2 license, provide users with unparalleled freedom to modify and improve the technology, ensuring it remains adaptable to their unique requirements.

“By leveraging open-source models, enterprises are not just passive consumers of technology; they become active contributors to a broader AI ecosystem,” Gil explained. This participatory approach accelerates innovation and ensures that AI advancements are grounded in real-world applications and challenges. The open-source community’s collaborative spirit also means that improvements and breakthroughs can be rapidly disseminated, benefiting all users.

Moreover, open-source frameworks offer a level of transparency and trust that is crucial in today’s data-driven world. Users can scrutinize the underlying code, understand the data used to train models and ensure compliance with regulatory and ethical standards. “Transparency is key to building trust in AI systems,” Gil emphasized. “When enterprises can see and understand what goes into their AI, they are more likely to embrace and deploy these technologies confidently.”

IBM’s commitment to open source is further exemplified by its contributions to major projects and partnerships within the community. The company’s involvement in the AI Alliance, launched in collaboration with Meta, brings together nearly 100 institutions, including leading universities, startups, and large-scale enterprises. This alliance aims to advance AI in a way that reflects the diversity and complexity of global societies, fostering inclusive and beneficial innovations for all.

In summary, embracing open source is not just a strategic choice for IBM; it is a fundamental philosophy that drives the company’s approach to AI. By championing open-source models and methodologies, IBM is positioning itself at the forefront of AI innovation, ensuring that the technology evolves in a way that is transparent, collaborative, and aligned with the needs of businesses and society. As Gil succinctly put it, “The future of AI is open, and together, we can build a more innovative and equitable world.”

Foundation Models: The Bedrock of AI

Foundation models have emerged as the cornerstone of modern AI, underpinning the transformative capabilities that are revolutionizing industries across the globe. In his keynote, Darío Gil underscored the significance of these models, emphasizing their role in encoding vast amounts of data and knowledge into highly capable AI systems. “The power of foundation models lies in their ability to represent and process data in previously unimaginable ways,” Gil noted. “They enable us to capture the complexity and nuance of human knowledge, making it accessible and actionable.”

One of the key advantages of foundation models is their scalability. These models can be trained on enormous datasets, incorporating a wide array of information from different domains. This scalability not only enhances their performance but also allows them to be applied to a variety of use cases. Gil highlighted IBM’s Granite family of models as a prime example, showcasing their versatility in handling tasks from natural language processing to coding and geospatial analysis. “These models are designed to be adaptable, ensuring that they can meet the diverse needs of enterprises,” he said.

The integration of multimodal data is another critical feature of foundation models. By combining information from text, images, audio, and other data types, these models can create richer and more accurate representations of the world. This capability is particularly valuable in applications such as autonomous vehicles, healthcare diagnostics, and financial analysis, where understanding the context and relationships between different data types is essential. “Multimodality is a game-changer,” Gil asserted. “It allows us to build AI systems that can understand and interact with the world in more sophisticated ways.”

Furthermore, foundation models are instrumental in democratizing AI. Providing a robust and flexible base enables organizations of all sizes to leverage advanced AI capabilities without requiring extensive in-house expertise. This democratization is facilitated by open-source initiatives, which make these powerful tools accessible to a broader audience. As exemplified by the Granite models, IBM’s commitment to open source ensures that AI’s benefits are widely shared, fostering innovation and inclusivity. “Open-source foundation models are leveling the playing field,” Gil remarked. “They empower companies to innovate and compete on a global scale.”

The potential of foundation models extends beyond current applications, promising to drive future advancements in AI. As these models evolve, they will unlock new possibilities and address increasingly complex challenges. Gil called on enterprises to actively engage in this evolution by contributing their data and expertise to enhance the models further. “The future of AI is a collaborative journey,” he said. “By working together, we can push the boundaries of what is possible and create AI systems that are more powerful, reliable, and beneficial for all.”

Foundation models represent a fundamental shift in AI technology, providing the bedrock upon which future innovations will be built. Their scalability, multimodal capabilities, and democratizing impact make them indispensable tools for enterprises seeking to harness the full potential of AI. As Gil eloquently put it, “Foundation models are not just technological advancements; they are enablers of a new era of human ingenuity and progress.”

A New Methodology: Instruct Lab

To revolutionize how enterprises interact with AI, IBM Research introduced a groundbreaking methodology called Instruct Lab. This innovative approach allows businesses to enhance their AI models incrementally, adding new skills and knowledge progressively, much like human learning. “Instruct Lab is a game-changer in the realm of AI development,” Darío Gil declared. “It enables us to teach AI in a more natural, human-like way, which is crucial for developing specialized capabilities efficiently.”

Instruct Lab stands out for its ability to integrate new information without starting from scratch, making the process both time and cost-efficient. Using a base model as a starting point, enterprises can introduce specific domain knowledge and skills, allowing the model to evolve and improve continuously. This approach contrasts sharply with traditional fine-tuning methods that often require creating multiple specialized models for different tasks. “With Instruct Lab, we can build upon a solid foundation, adding layers of expertise without losing the generality and robustness of the original model,” Gil explained.

One of the key features of Instruct Lab is its use of a teacher model to generate synthetic data, which is then used to train the AI. This process ensures that the model can learn from a broad range of examples, enhancing its ability to understand and respond to various scenarios. “Synthetic data generation is a powerful tool in our methodology,” Gil noted. “It allows us to scale the training process efficiently, providing the model with the diversity of experiences needed to perform well in real-world applications.”

The methodology also emphasizes transparency and control, ensuring that enterprises have full visibility into the training process and the data being used. This transparency is crucial for maintaining trust and ensuring the security of enterprise data. “Instruct Lab is designed with enterprise needs in mind,” Gil emphasized. “We prioritize transparency and control, allowing businesses to understand and trust the AI systems they are developing.”

The impact of the Instruct Lab is already evident in IBM’s own projects. For instance, the development of the IBM Watson X Code Assistant for Z demonstrated the methodology’s effectiveness. By applying Instruct Lab, IBM was able to significantly enhance the model’s understanding of COBOL, a critical language for mainframe applications. “In just one week, we achieved results that surpassed months of traditional fine-tuning,” Gil shared. “This showcases the incredible potential of Instruct Lab to accelerate AI development and deliver superior performance.”

The introduction of Instruct Lab represents a significant step forward in AI technology, providing enterprises with a robust and flexible tool for continuous improvement. As businesses increasingly rely on AI to drive innovation and efficiency, methodologies like Instruct Lab will be essential for staying ahead of the curve. “Instruct Lab embodies our commitment to empowering enterprises with cutting-edge AI capabilities,” Gil concluded. “It is a testament to our dedication to advancing AI in ways that are both practical and transformative.”

Scaling AI in Enterprises

Scaling AI in enterprises is not just about deploying advanced algorithms; it’s about integrating these technologies seamlessly into the fabric of the business to drive meaningful impact. Darío Gil emphasized the transformative potential of AI when it’s scaled correctly within enterprises. “The real power of AI comes from its ability to enhance every aspect of an organization,” he stated. “From optimizing supply chains to personalizing customer interactions, the possibilities are limitless when AI is effectively scaled.”

One of the critical challenges in scaling AI is ensuring that the technology is accessible and usable across various departments and functions within an organization. IBM’s approach addresses this by providing robust tools and frameworks that allow businesses to customize AI models to their specific needs. “We recognize that every enterprise has unique requirements,” Gil noted. “Our solutions are designed to be flexible and adaptable, enabling companies to tailor AI to their particular contexts and goals.”

Moreover, scaling AI requires a strong foundation of data management and governance. Enterprises must be able to trust the data that feeds their AI models, ensuring it is accurate, secure, and used ethically. IBM places a strong emphasis on data governance as a cornerstone of its AI strategy. “Data is the lifeblood of AI,” Gil explained. “Without proper governance and management, the insights derived from AI could be flawed. We provide comprehensive tools to help enterprises manage their data effectively, ensuring that their AI initiatives are built on a solid foundation.”

To truly scale AI, enterprises must also invest in the continuous training and development of their workforce. AI is not a set-it-and-forget-it solution; it requires ongoing learning and adaptation. IBM supports this through its extensive training programs and resources, helping organizations develop the skills needed to harness the full potential of AI. “Human expertise is essential in driving AI success,” Gil said. “We are committed to empowering our clients with the knowledge and skills they need to excel in an AI-driven world.”

Additionally, IBM’s focus on open-source models plays a crucial role in scaling AI. By leveraging open-source technologies, enterprises can benefit from a collaborative approach to AI development, accessing a wealth of community-driven innovations and best practices. “The open-source community is a vital component of AI advancement,” Gil highlighted. “It fosters a spirit of collaboration and continuous improvement, essential for scaling AI effectively across enterprises.”

As enterprises navigate the complexities of scaling AI, IBM’s comprehensive approach—spanning advanced technologies, robust data management, continuous learning, and open-source collaboration—provides a clear pathway to success. “Scaling AI is a journey,” Gil concluded. “It’s about creating a sustainable, adaptable framework that grows with the enterprise, driving innovation and competitive advantage at every step.”

Looking Ahead

As IBM continues to push the boundaries of AI, the future holds immense potential for enterprises willing to embrace these transformative technologies. Darío Gil’s vision for AI is one where innovation and collaboration drive progress, ensuring that AI serves not just as a tool for efficiency but as a catalyst for groundbreaking advancements across industries.

One of the key areas of focus for IBM moving forward is the integration of AI with other cutting-edge technologies, such as quantum computing and blockchain. “The convergence of AI with quantum computing can unlock new levels of problem-solving capabilities that were previously unimaginable,” Gil noted. “By combining the strengths of these technologies, we can tackle some of the most complex challenges facing humanity, from climate change to healthcare.”

IBM is also committed to ensuring that AI development remains ethical and inclusive. The company is actively working on initiatives to address biases in AI models and to promote transparency and accountability in AI systems. “As we look ahead, it’s crucial that we build AI that is fair, transparent, and respects the values of our society,” Gil emphasized. “We are dedicated to leading the charge in creating ethical AI frameworks that benefit everyone.”

In enterprise applications, IBM plans to expand its portfolio of AI-driven solutions, providing businesses with even more tools to enhance their operations and drive innovation. The company’s continued investment in research and development ensures its clients have access to the latest advancements in AI technology. “Our goal is to empower enterprises to leverage AI in ways that were previously thought impossible,” Gil said. “We are constantly exploring new frontiers and developing solutions that will keep our clients at the forefront of their industries.”

Moreover, IBM’s commitment to open-source AI models will play a significant role in the future of AI development. By fostering a collaborative environment, IBM aims to accelerate the pace of innovation and ensure that AI technology evolves in a way that is beneficial for all stakeholders. “The future of AI is one that is built on collaboration and shared knowledge,” Gil stated. “By embracing open-source principles, we can create a thriving ecosystem where everyone has the opportunity to contribute and benefit from AI advancements.”

As the landscape of AI continues to evolve, IBM remains steadfast in its mission to drive technological progress while addressing its ethical and societal implications. “The road ahead is full of exciting possibilities,” Gil concluded. We are committed to leading the way in AI innovation, ensuring that our advancements serve the greater good and pave the way for a better future for all.”

With a forward-looking approach that combines technological excellence, ethical considerations, and a collaborative spirit, IBM is well-positioned to shape the future of AI and drive meaningful change across the globe. As enterprises prepare to navigate this dynamic landscape, they can look to IBM for guidance, support, and innovative solutions to help them thrive in the age of AI.

]]>
604889