According to Bloomberg, Khan made the comments at Y Combinator in San Francisco.
“There’s tremendous potential for open-weight models to promote competition,” Khan said. “Open-weight models can liberate startups from the arbitrary whims of closed developers and cloud gatekeepers.”
Khan’s comments come at a time when regulators on both sides of the Atlantic are growing increasingly wary of Big Tech. AI companies have done little to stave off such concerns, with accusations they plagiarize content, throttle organizations’ servers as they scrape them, and show little regard for the potential danger AI may pose.
In view those issues, many lawmakers are concerned about a future where AI development and breakthroughs are largely controlled by a handful of companies.
One notable exception in the industry is Meta’s Llama AI model, which the company has made available as open-source software. The company explained its reasons in a blog post announcing Llama 3:
We’re committed to the continued growth and development of an open AI ecosystem for releasing our models responsibly. We have long believed that openness leads to better, safer products, faster innovation, and a healthier overall market. This is good for Meta, and it is good for society. We’re taking a community-first approach with Llama 3, and starting today, these models are available on the leading cloud, hosting, and hardware platforms with many more to come.
With Khan’s comments, LLama and other open models may see an uptick in use.
]]>The White House announced the news in a press release:
Nine months ago, President Biden issued a landmark Executive Order to ensure that America leads the way in seizing the promise and managing the risks of artificial intelligence (AI).
This Executive Order built on the voluntary commitments he and Vice President Harris received from 15 leading U.S. AI companies last year. Today, the administration announced that Apple has signed onto the voluntary commitments, further cementing these commitments as cornerstones of responsible AI innovation.
Apple is widely considered to be a significant factor in the AI industry, thanks largely to its penchant for making high-tech solutions approach to the average user, as well as the huge user base that it can leverage.
With the announcement of Apple Intelligence, many critics and experts say Apple has done more to make the case for AI’s usefulness to the average user than most other companies combined. In view of the role Apple will likely play, it’s good to see the company’s continued commitment to safe AI development and deployment.
]]>Musk’s xAI originally tapped Oracle to help it deploy 24,000 H100s that were used to train its Grok 2 model. According to Musk, however, the company plans to go its own way, building out its own cluster containing some 100,000 H100s. Musk framed the decision in the context of needing to leapfrog its AI rivals, with controlling its own cluster being the key to doing so.
xAI contracted for 24k H100s from Oracle and Grok 2 trained on those. Grok 2 is going through finetuning and bug fixes. Probably ready to release next month.
xAI is building the 100k H100 system itself for fastest time to completion. Aiming to begin training later this month. It will be the most powerful training cluster in the world by a large margin.
The reason we decided to do the 100k H100 and next major system internally was that our fundamental competitiveness depends on being faster than any other AI company. This is the only way to catch up.
Oracle is a great company and there is another company that shows promise also involved in that OpenAI GB200 cluster, but, when our fate depends on being the fastest by far, we must have our own hands on the steering wheel, rather than be a backseat driver.
Elon Musk (@elonmusk) | July 9, 2024
The move is a blow to Oracle. As Investors.com points out, Oracle founder Larry Ellison touted its relationship with xAI in a recent quarterly earnings call, saying his company was working to secure more H100s for the startup.
“We gave them quite a few,” Ellison said at the time. “But they wanted more, and we are in the process of getting them more.”
]]>Prompts are an important part of the AI development process, and can have a major impact on the results, as Anthropic says in a blog post announcing the new feature:
When building AI-powered applications, prompt quality significantly impacts results. But crafting high quality prompts is challenging, requiring deep knowledge of your application’s needs and expertise with large language models. To speed up development and improve outcomes, we’ve streamlined this process to make it easier for users to produce high quality prompts.
You can now generate, test, and evaluate your prompts in the Anthropic Console. We’ve added new features, including the ability to generate automatic test cases and compare outputs, that allow you to leverage Claude to generate the very best responses for your needs.
Anthropic says users can generate prompts simply by describing a task to Claude. Using the Claude 3.5 Sonnet engine, Claude will use the description its given to generate a high-quality prompt.
The new Evaluate feature makes it much easier to test prompts against real-world inputs.
Testing prompts against a range of real-world inputs can help you build confidence in the quality of your prompt before deploying it to production. With the new Evaluate feature you can do this directly in our Console instead of manually managing tests across spreadsheets or code.
Manually add or import new test cases from a CSV, or ask Claude to auto-generate test cases for you with the ‘Generate Test Case’ feature. Modify your test cases as needed, then run all of the test cases in one click. View and adjust Claude’s understanding of the generation requirements for each variable to get finer-grained control over the test cases Claude generates.
Anthropic is already the leading OpenAI competitor, with its Claude 3.5 besting OpenAI’s GPT-4o in a range of tests. With the new features aimed at improving the quality of prompts, Anthropic continues to push AI development forward.
OpenAI announced in late June that it would block API traffic from countries that were not on its “supported countries and territories” list. Users on the company’s forums reported receiving emails from the company informing them of the policy.
China was conspicuously absent from the long list of supported countries, meaning that Chinese developers will not have access to the company’s API for development. According to The Information, however, there is a significant workaround that runs straight through Microsoft Azure AI.
Developers in China who want to take advantage of OpenAI’s models can still do so if they sign up for an Azure account. Because of Microsoft and OpenAI’s close relationship, this gives developers access to the AI firm’s AI models through Microsoft’s services.
According to the outlet, the exception works because Azure China is a joint venture with Chinese company 21Vianet. Multiple customers confirmed to The Information that they had full access to OpenAI models within Azure.
Given the importance of the Chinese market, the revelation is good news for Microsoft, OpenAI, and Chinese AI developers.
]]>AI models are designed to operate within strictly defined parameters that ensure the responses it gives are not offensive and do not cause harm. This is something AI firms have struggled with, with AI models sometimes going beyond their parameters and stirring up controversy in the process.
According to Microsoft Security, there is a newly discovered jailbreak attack—Skeleton Key—that impacts multiple AI models from various firms (hence the name).
This AI jailbreak technique works by using a multi-turn (or multiple step) strategy to cause a model to ignore its guardrails. Once guardrails are ignored, a model will not be able to determine malicious or unsanctioned requests from any other. Because of its full bypass abilities, we have named this jailbreak technique Skeleton Key.
This threat is in the jailbreak category, and therefore relies on the attacker already having legitimate access to the AI model. In bypassing safeguards, Skeleton Key allows the user to cause the model to produce ordinarily forbidden behaviors, which could range from production of harmful content to overriding its usual decision-making rules. Like all jailbreaks, the impact can be understood as narrowing the gap between what the model is capable of doing (given the user credentials, etc.) and what it is willing to do. As this is an attack on the model itself, it does not impute other risks on the AI system, such as permitting access to another user’s data, taking control of the system, or exfiltrating data.
Microsoft says it has already made a number of updates to its Copilot AI assistants and other LLM technology in an effort to mitigate the attack. The company says customers should consider the following actions to implement their own AI system design:
- Input filtering: Azure AI Content Safety detects and blocks inputs that contain harmful or malicious intent leading to a jailbreak attack that could circumvent safeguards.
- System message: Prompt engineering the system prompts to clearly instruct the large language model (LLM) on appropriate behavior and to provide additional safeguards. For instance, specify that any attempts to undermine the safety guardrail instructions should be prevented (read our guidance on building a system message framework here).
- Output filtering: Azure AI Content Safety post-processing filter that identifies and prevents output generated by the model that breaches safety criteria.
- Abuse monitoring: Deploying an AI-driven detection system trained on adversarial examples, and using content classification, abuse pattern capture, and other methods to detect and mitigate instances of recurring content and/or behaviors that suggest use of the service in a manner that may violate guardrails. As a separate AI system, it avoids being influenced by malicious instructions. Microsoft Azure OpenAI Service abuse monitoring is an example of this approach.
The company says its Azure AI tools already help customers protect against this type of attack as well:
Microsoft provides tools for customers developing their own applications on Azure. Azure AI Content Safety Prompt Shields are enabled by default for models hosted in the Azure AI model catalog as a service, and they are parameterized by a severity threshold. We recommend setting the most restrictive threshold to ensure the best protection against safety violations. These input and output filters act as a general defense not only against this particular jailbreak technique, but also a broad set of emerging techniques that attempt to generate harmful content. Azure also provides built-in tooling for model selection, prompt engineering, evaluation, and monitoring. For example, risk and safety evaluations in Azure AI Studio can assess a model and/or application for susceptibility to jailbreak attacks using synthetic adversarial datasets, while Microsoft Defender for Cloud can alert security operations teams to jailbreaks and other active threats.
With the integration of Azure AI and Microsoft Security (Microsoft Purview and Microsoft Defender for Cloud) security teams can also discover, protect, and govern these attacks. The new native integration of Microsoft Defender for Cloud with Azure OpenAI Service, enables contextual and actionable security alerts, driven by Azure AI Content Safety Prompt Shields and Microsoft Defender Threat Intelligence. Threat protection for AI workloads allows security teams to monitor their Azure OpenAI powered applications in runtime for malicious activity associated with direct and in-direct prompt injection attacks, sensitive data leaks and data poisoning, or denial of service attacks.
The Skeleton Key attack underscores the ongoing challenges facing companies as AI becomes more widely used. While it can be a valuable tool for cybersecurity, it can also open up entirely new attack vectors.
]]>Properly evaluating AI’s potential is a growing challenge for AI firms as the technology evolves. Not only is it challenging to properly evaluate an AI’s capabilities, but there are also concerns with evaluating the various safety issues involved.
Anthropic has increasingly been setting itself apart in the AI field, not only for its powerful Claude model that is currently beating OpenAI’s GPT-4o, but also for its safety-first approach to AI. In fact, the company was founded by OpenAI executives that were concerned with the direction of OpenAI, and the company has continued to attract disillusioned OpenAI engineers. The most notable recent example is Jan Leike, who left OpenAI after the safety team he co-lead was disbanded.
With that background, it’s not surprising that Anthropic is interested in developing and discovering new and better ways to properly evaluate AI. The company outlines its highest priority areas of focus:
- AI Safety Level assessments
- Advanced capability and safety metrics
- Infrastructure, tools, and methods for developing evaluations
The company outlines a number of AI Safety Levels (ASLs) that is concerned with, including cybersecurity; chemical, biological, radiological, and nuclear (CBRN) risks; model autonomy; national security risks; and misalignment risks. In all of these areas, the company is concerned with the risk that AI could be used to aid individuals in doing harm.
We’re particularly interested in capabilities that, if automated and scaled, could pose significant risks to critical infrastructure and economically valuable systems at levels approaching advanced persistent threat actors.
We’re prioritizing evaluations that assess two critical capabilities: a) the potential for models to significantly enhance the abilities of non-experts or experts in creating CBRN threats, and b) the capacity to design novel, more harmful CBRN threats.
AI systems have the potential to significantly impact national security, defense, and intelligence operations of both state and non-state actors. We’re committed to developing an early warning system to identify and assess these complex emerging risks.
Anthropic reveals a fascinating, and terrifying, observation about current AI models, what the company identifies as “misalignment risks.”
Our research shows that, under some circumstances, AI models can learn dangerous goals and motivations, retain them even after safety training, and deceive human users about actions taken in their pursuit.
The company says this represents a major danger moving forward as AI models become more advanced.
These abilities, in combination with the human-level persuasiveness and cyber capabilities of current AI models, increases our concern about the potential actions of future, more-capable models. For example, future models might be able to pursue sophisticated and hard-to-detect deception that bypasses or sabotages the security of an organization, either by causing humans to take actions they would not otherwise take or exfiltrating sensitive information.
Anthropic goes on to highlight its desire to improve evaluation methods to address bias issues, something that has been a significant challenge in training existing AI models.
Evaluations that provide sophisticated, nuanced assessments that go beyond surface-level metrics to create rigorous assessments targeting concepts like harmful biases, discrimination, over-reliance, dependence, attachment, psychological influence, economic impacts, homogenization, and other broad societal impacts.
The company also wants to ensure AI benchmarks support multiple languages, something that is not currently the case. New evaluation methods should also be able to “detect potentially harmful model outputs,” such as “attempts to automate cyber incidents.” The company also wants the new evaluation methods to better determine AI’s ability to learn, especially in the sciences.
Parties interested in submitting a proposal should keep the company’s 10 requirements in mind:
- Sufficiently difficult: Evaluations should be relevant for measuring the capabilities listed for levels ASL-3 or ASL-4 in our Responsible Scaling Policy, and/or human-expert level behavior.
- Not in the training data: Too often, evaluations end up measuring model memorization because the data is in its training set. Where possible and useful, make sure the model hasn’t seen the evaluation. This helps indicate that the evaluation is capturing behavior that generalizes beyond the training data.
- Efficient, scalable, ready-to-use: Evaluations should be optimized for efficient execution, leveraging automation where possible. They should be easily deployable using existing infrastructure with minimal setup.
- High volume where possible: All else equal, evaluations with 1,000 or 10,000 tasks or questions are preferable to those with 100. However, high-quality, low-volume evaluations are also valuable.
- Domain expertise: If the evaluation is about expert performance on a particular subject matter (e.g. science), make sure to use subject matter experts to develop or review the evaluation.
- Diversity of formats: Consider using formats that go beyond multiple choice, such as task-based evaluations (for example, seeing if code passes a test or a flag is captured in a CTF), model-graded evaluations, or human trials.
- Expert baselines for comparison: It is often useful to compare the model’s performance to the performance of human experts on that domain.
- Good documentation and reproducibility: We recommend documenting exactly how the evaluation was developed and any limitations or pitfalls it is likely to have. Use standards like the Inspect or the METR standard where possible.
- Start small, iterate, and scale: Start by writing just one to five questions or tasks, run a model on the evaluation, and read the model transcripts. Frequently, you’ll realize the evaluation doesn’t capture what you want to test, or it’s too easy.
- Realistic, safety-relevant threat modeling: Safety evaluations should ideally have the property that if a model scored highly, experts would believe that a major incident could be caused. Most of the time, when models have performed highly, experts have realized that high performance on that version of the evaluation is not sufficient to worry them.
Those interested in submitting a proposal, and possibly working long-term with Anthropic, should use this application form.
OpenAI has been criticized for for a lack of transparency that has led many to believe the company has lost its way and is no longer focused on its one-time goal of safe AI development. Anthropic’s willingness to engage the community and industry is a refreshing change of pace.
]]>Labeling AI content has become a growing concern for online platforms, as well as regulators, as AI-generated content has become so realistic that it could easily be used to create false narratives. Meta announced in April plans to label AI content with a “Made with AI” label. Unfortunately, it’s algorithm for identifying AI content had some issues, with photos taken by human photographers being improperly labeled.
The company says it has made changes to address the issue.
We want people to know when they see posts that have been made with AI. Earlier this year, we announced a new approach for labeling AI-generated content. An important part of this approach relies on industry standard indicators that other companies include in content created using their tools, which help us assess whether something is created using AI.
Like others across the industry, we’ve found that our labels based on these indicators weren’t always aligned with people’s expectations and didn’t always provide enough context. For example, some content that included minor modifications using AI, such as retouching tools, included industry standard indicators that were then labeled “Made with AI.” While we work with companies across the industry to improve the process so our labeling approach better matches our intent, we’re updating the “Made with AI” label to “AI info” across our apps, which people can click for more information.
According to CNET, photographer Pete Souza said cropping tools appear to be one of the culprits. Because such tools add information to images, it seems that Meta’s algorithm was incorrectly identifying that added information and taking it as an indication the images were AI-generated.
The entire issue demonstrates the growing challenges associated with correctly identifying AI-generated content. For years, experts have warned about the potential havoc deepfakes could cause, impacting everything from people’s personal lives to business to politics.
Interestingly, OpenAI shuttered its own AI-content detection tool in early 2024, saying at the time that such tools don’t work:
While some (including OpenAI) have released tools that purport to detect AI-generated content, none of these have proven to reliably distinguish between AI-generated and human-generated content.
It remains to be seen if Meta will be able to reliably identify AI-generated images, or if it will suffer the same issues that led OpenAI to throw in the towel.
]]>Content ownership has become one of the most controversial and contentious aspects of AI development. AI models need vast amounts of data for training and most AI firms have used online data as their source, some establishing paid content deals, and others scraping data without paying for it.
According to Suleyman, content found online should be governed by fair use.
“I think that with respect to content that is already on the open web, the social contract of that content since the 1990s has been it is fair use,” he said in an interview with CNBC, via The Register. “Anyone can copy it, recreate with it, reproduce with it. That has been freeware, if you like. That’s been the understanding.”
The one exception Suleyman made is for websites and publishers that explicitly forbid content scraping.
“There’s a separate category where a website or publisher or news organization had explicitly said, ‘do not scrape or crawl me for any other reason than indexing me,’ so that other people can find that content,” he explained. “But that’s the gray area. And I think that’s going to work its way through the courts.”
Unfortunately, not all AI firms respect ‘do not crawl’ requests. AWS is investigating Perplexity AI over accusations it is scraping websites in violation of their terms, using AWS’ cloud platform to do so.
As Suleyman points out, the legality of the practice will ultimately be decided in the courts but, in the meantime, content ownership will continue to be a hotly debated topic.
Content ownership is one of the biggest legal and ethical challenges facing AI firms. Some AI firms have committed to honoring the Robots Exclusion Protocol, a standard that defines what web pages search engines should index or scrape, and which should be ignored.
Perplexity AI has been accused of ignoring the protocol and scraping sites without permission, according to Wired. As a result, sources within AWS confirmed to the outlet that it was investigating the AI firm, since AWS requires that its cloud customers adhere to the Robots Exclusion Protocol.
“AWS’s terms of service prohibit customers from using our services for any illegal activity, and our customers are responsible for complying with our terms and all applicable laws,” the spokesperson said in a statement to Wired.
Perplexity has already faced accusations of stealing articles and plagiarism. If the AWS investigation goes against the AI firm, it remains to be seen what action the cloud provider may take.
]]>Reddit is a popular place for AI companies to scrape, thanks to the large quantity of user-generated content on a vast array of subjects. Reddit has signed a deal with Google allowing the company to use the site’s content, but other companies appear to be continuing their efforts to scrape the site.
The company says it will make changes to address the issue.
In the coming weeks, we’ll update our Robots Exclusion Protocol (robots.txt file), which gives high-level instructions about how we do and don’t allow Reddit to be crawled by third parties. Along with our updated robots.txt file, we will continue rate-limiting and/or blocking unknown bots and crawlers from accessing reddit.com. This update shouldn’t impact the vast majority of folks who use and enjoy Reddit. Good faith actors – like researchers and organizations such as the Internet Archive – will continue to have access to Reddit content for non-commercial use.
Mark Graham, Director, Wayback Machine at Internet Archive, praised Reddit’s position.
“The Internet Archive is grateful that Reddit appreciates the importance of helping to ensure the digital records of our times are archived and preserved for future generations to enjoy and learn from,” said Graham. “Working in collaboration with Reddit we will continue to record and make available archives of Reddit, along with the hundreds of millions of URLs from other sites we archive every day.”
Reddit emphasized that organizations must abide by its policies.
]]>Anyone accessing Reddit content must abide by our policies, including those in place to protect redditors. We are selective about who we work with and trust with large-scale access to Reddit content. Organizations looking to access Reddit content can head over to our guide to accessing Reddit Data.
AI models are growing in popularity, but not all users are comfortable giving the companies behind them access to their data. LibreChat is an open-source AI platform that gives user access to multiple models in a centralized hub.
LibreChat is a free, open-source AI chat platform that empowers you to harness the capabilities of cutting-edge language models from multiple providers in a unified interface. With its vast customization options, innovative enhancements, and seamless integration of AI services, LibreChat offers an unparalleled conversational experience.
LibreChat is an enhanced, open-source ChatGPT clone that brings together the latest advancements in AI technology. It serves as a centralized hub for all your AI conversations, providing a familiar, user-friendly interface enriched with advanced features and customization capabilities.
The project emphasizes several design principles, not the least of which are privacy and security.
- User-Friendly Interface: Inspired by the familiar ChatGPT UI, LibreChat offers a clean and intuitive layout, making it easy for users to engage with AI assistants.
- Multimodal Conversations: LibreChat supports multimodal conversations, allowing you to upload and analyze images, chat with files, and leverage advanced agent capabilities powered by AI models like GPT-4 Claude and Gemini Vision.
- Extensibility: With its plugin architecture and open-source nature, LibreChat encourages the development of custom extensions and integrations, enabling users to tailor the platform to their specific needs.
- Privacy and Security: LibreChat prioritizes privacy and security by offering secure authentication, moderation tools, and the ability to self-host the application.
LibreChat is an intriguing entrance in the AI market, one that demonstrates the ingenuity of the open-source community.
]]>This evolution in recruitment strategies invites companies to leverage cutting-edge tools on a tech recruitment platform, such as a Huntly, and stay ahead in the talent acquisition game.
AI is revolutionizing the initial stages of recruitment by automating the screening of candidates. Machine learning algorithms can sift through hundreds of applications, identifying those that best match the job requirements based on skills, experience, and potential. This not only saves valuable time but also minimizes human bias, ensuring a more diverse and qualified candidate pool progresses to the interview stage.
Beyond screening, AI systems are becoming increasingly sophisticated at matching candidates with job vacancies. By analyzing data points across previous successful hires and ongoing performance metrics, these systems can predict candidate success more accurately. This approach not only improves the quality of hires but also contributes to longer-term employee satisfaction and retention.
Automation is streamlining administrative tasks such as interview scheduling and candidate communication. Chatbots and AI-driven platforms can handle queries, provide updates, and manage scheduling without human intervention, enhancing the candidate experience by ensuring prompt and personalized interaction throughout the recruitment process.
The rise of virtual interviews and assessment tools is making geographical boundaries irrelevant. Online coding tests, virtual reality (VR) environments for real-world problem-solving, and video interviews allow for a comprehensive evaluation of candidates’ technical and soft skills without the need for physical presence. This opens up a global talent pool, enabling companies to attract and assess candidates from across the world.
Predictive analytics is shaping the future of recruitment by forecasting hiring needs and candidate success. By analyzing trends, skills evolution, and company growth patterns, businesses can anticipate future recruitment needs and build strategic talent pipelines. This proactive approach ensures companies are always prepared to meet their developmental and technological challenges with the right talent.
Recognizing the fast-paced nature of technological advancement, recruitment strategies are increasingly focusing on candidates’ potential for continuous learning. Platforms that offer ongoing skill assessment and development opportunities are becoming integral to the recruitment process, allowing companies to not only hire for current needs but also invest in the future growth of their employees.
As technology reshapes recruitment, ethical considerations, and human oversight remain paramount. Ensuring AI and automation are used responsibly to avoid biases and protect candidate privacy is crucial. Companies must balance technological efficiency with a human touch, ensuring that recruitment processes remain fair, transparent, and respectful of candidates’ rights.
The future of developer recruitment is intricately linked to AI, automation, and technological innovations. These advancements promise a more efficient, accurate, and global recruitment process, allowing companies to meet the growing demand for tech talent effectively.
However, navigating this future requires a mindful approach to leveraging technology while maintaining ethical standards and human connections. As the industry continues to evolve, embracing these innovations while staying grounded in human-centric recruitment practices will be key to attracting and retaining the best tech talent.
]]>“The future of AI is open,” declared Gil, emphasizing the importance of open-source innovation and collaborative efforts in the AI landscape. He urged businesses to adopt open strategies to maximize the potential of their AI systems, arguing that such an approach not only fosters innovation but also ensures flexibility and adaptability.
“Open is about innovating together, not in isolation,” Gil said. By choosing open-source frameworks, companies can decide which models to use, what data to integrate, and how to adapt AI to their specific needs. Gil argued that this collaborative approach is essential for the evolution of AI to meet the diverse aspirations of various industries.
The strength of open source lies in its ability to foster a community-driven ecosystem where innovation can thrive unencumbered by proprietary constraints. Gil pointed to the success of IBM’s own Granite family of models, designed to handle tasks ranging from coding to time series analysis and geospatial data processing. These models, released under an Apache 2 license, provide users with unparalleled freedom to modify and improve the technology, ensuring it remains adaptable to their unique requirements.
“By leveraging open-source models, enterprises are not just passive consumers of technology; they become active contributors to a broader AI ecosystem,” Gil explained. This participatory approach accelerates innovation and ensures that AI advancements are grounded in real-world applications and challenges. The open-source community’s collaborative spirit also means that improvements and breakthroughs can be rapidly disseminated, benefiting all users.
Moreover, open-source frameworks offer a level of transparency and trust that is crucial in today’s data-driven world. Users can scrutinize the underlying code, understand the data used to train models and ensure compliance with regulatory and ethical standards. “Transparency is key to building trust in AI systems,” Gil emphasized. “When enterprises can see and understand what goes into their AI, they are more likely to embrace and deploy these technologies confidently.”
IBM’s commitment to open source is further exemplified by its contributions to major projects and partnerships within the community. The company’s involvement in the AI Alliance, launched in collaboration with Meta, brings together nearly 100 institutions, including leading universities, startups, and large-scale enterprises. This alliance aims to advance AI in a way that reflects the diversity and complexity of global societies, fostering inclusive and beneficial innovations for all.
In summary, embracing open source is not just a strategic choice for IBM; it is a fundamental philosophy that drives the company’s approach to AI. By championing open-source models and methodologies, IBM is positioning itself at the forefront of AI innovation, ensuring that the technology evolves in a way that is transparent, collaborative, and aligned with the needs of businesses and society. As Gil succinctly put it, “The future of AI is open, and together, we can build a more innovative and equitable world.”
Foundation models have emerged as the cornerstone of modern AI, underpinning the transformative capabilities that are revolutionizing industries across the globe. In his keynote, Darío Gil underscored the significance of these models, emphasizing their role in encoding vast amounts of data and knowledge into highly capable AI systems. “The power of foundation models lies in their ability to represent and process data in previously unimaginable ways,” Gil noted. “They enable us to capture the complexity and nuance of human knowledge, making it accessible and actionable.”
One of the key advantages of foundation models is their scalability. These models can be trained on enormous datasets, incorporating a wide array of information from different domains. This scalability not only enhances their performance but also allows them to be applied to a variety of use cases. Gil highlighted IBM’s Granite family of models as a prime example, showcasing their versatility in handling tasks from natural language processing to coding and geospatial analysis. “These models are designed to be adaptable, ensuring that they can meet the diverse needs of enterprises,” he said.
The integration of multimodal data is another critical feature of foundation models. By combining information from text, images, audio, and other data types, these models can create richer and more accurate representations of the world. This capability is particularly valuable in applications such as autonomous vehicles, healthcare diagnostics, and financial analysis, where understanding the context and relationships between different data types is essential. “Multimodality is a game-changer,” Gil asserted. “It allows us to build AI systems that can understand and interact with the world in more sophisticated ways.”
Furthermore, foundation models are instrumental in democratizing AI. Providing a robust and flexible base enables organizations of all sizes to leverage advanced AI capabilities without requiring extensive in-house expertise. This democratization is facilitated by open-source initiatives, which make these powerful tools accessible to a broader audience. As exemplified by the Granite models, IBM’s commitment to open source ensures that AI’s benefits are widely shared, fostering innovation and inclusivity. “Open-source foundation models are leveling the playing field,” Gil remarked. “They empower companies to innovate and compete on a global scale.”
The potential of foundation models extends beyond current applications, promising to drive future advancements in AI. As these models evolve, they will unlock new possibilities and address increasingly complex challenges. Gil called on enterprises to actively engage in this evolution by contributing their data and expertise to enhance the models further. “The future of AI is a collaborative journey,” he said. “By working together, we can push the boundaries of what is possible and create AI systems that are more powerful, reliable, and beneficial for all.”
Foundation models represent a fundamental shift in AI technology, providing the bedrock upon which future innovations will be built. Their scalability, multimodal capabilities, and democratizing impact make them indispensable tools for enterprises seeking to harness the full potential of AI. As Gil eloquently put it, “Foundation models are not just technological advancements; they are enablers of a new era of human ingenuity and progress.”
To revolutionize how enterprises interact with AI, IBM Research introduced a groundbreaking methodology called Instruct Lab. This innovative approach allows businesses to enhance their AI models incrementally, adding new skills and knowledge progressively, much like human learning. “Instruct Lab is a game-changer in the realm of AI development,” Darío Gil declared. “It enables us to teach AI in a more natural, human-like way, which is crucial for developing specialized capabilities efficiently.”
Instruct Lab stands out for its ability to integrate new information without starting from scratch, making the process both time and cost-efficient. Using a base model as a starting point, enterprises can introduce specific domain knowledge and skills, allowing the model to evolve and improve continuously. This approach contrasts sharply with traditional fine-tuning methods that often require creating multiple specialized models for different tasks. “With Instruct Lab, we can build upon a solid foundation, adding layers of expertise without losing the generality and robustness of the original model,” Gil explained.
One of the key features of Instruct Lab is its use of a teacher model to generate synthetic data, which is then used to train the AI. This process ensures that the model can learn from a broad range of examples, enhancing its ability to understand and respond to various scenarios. “Synthetic data generation is a powerful tool in our methodology,” Gil noted. “It allows us to scale the training process efficiently, providing the model with the diversity of experiences needed to perform well in real-world applications.”
The methodology also emphasizes transparency and control, ensuring that enterprises have full visibility into the training process and the data being used. This transparency is crucial for maintaining trust and ensuring the security of enterprise data. “Instruct Lab is designed with enterprise needs in mind,” Gil emphasized. “We prioritize transparency and control, allowing businesses to understand and trust the AI systems they are developing.”
The impact of the Instruct Lab is already evident in IBM’s own projects. For instance, the development of the IBM Watson X Code Assistant for Z demonstrated the methodology’s effectiveness. By applying Instruct Lab, IBM was able to significantly enhance the model’s understanding of COBOL, a critical language for mainframe applications. “In just one week, we achieved results that surpassed months of traditional fine-tuning,” Gil shared. “This showcases the incredible potential of Instruct Lab to accelerate AI development and deliver superior performance.”
The introduction of Instruct Lab represents a significant step forward in AI technology, providing enterprises with a robust and flexible tool for continuous improvement. As businesses increasingly rely on AI to drive innovation and efficiency, methodologies like Instruct Lab will be essential for staying ahead of the curve. “Instruct Lab embodies our commitment to empowering enterprises with cutting-edge AI capabilities,” Gil concluded. “It is a testament to our dedication to advancing AI in ways that are both practical and transformative.”
Scaling AI in enterprises is not just about deploying advanced algorithms; it’s about integrating these technologies seamlessly into the fabric of the business to drive meaningful impact. Darío Gil emphasized the transformative potential of AI when it’s scaled correctly within enterprises. “The real power of AI comes from its ability to enhance every aspect of an organization,” he stated. “From optimizing supply chains to personalizing customer interactions, the possibilities are limitless when AI is effectively scaled.”
One of the critical challenges in scaling AI is ensuring that the technology is accessible and usable across various departments and functions within an organization. IBM’s approach addresses this by providing robust tools and frameworks that allow businesses to customize AI models to their specific needs. “We recognize that every enterprise has unique requirements,” Gil noted. “Our solutions are designed to be flexible and adaptable, enabling companies to tailor AI to their particular contexts and goals.”
Moreover, scaling AI requires a strong foundation of data management and governance. Enterprises must be able to trust the data that feeds their AI models, ensuring it is accurate, secure, and used ethically. IBM places a strong emphasis on data governance as a cornerstone of its AI strategy. “Data is the lifeblood of AI,” Gil explained. “Without proper governance and management, the insights derived from AI could be flawed. We provide comprehensive tools to help enterprises manage their data effectively, ensuring that their AI initiatives are built on a solid foundation.”
To truly scale AI, enterprises must also invest in the continuous training and development of their workforce. AI is not a set-it-and-forget-it solution; it requires ongoing learning and adaptation. IBM supports this through its extensive training programs and resources, helping organizations develop the skills needed to harness the full potential of AI. “Human expertise is essential in driving AI success,” Gil said. “We are committed to empowering our clients with the knowledge and skills they need to excel in an AI-driven world.”
Additionally, IBM’s focus on open-source models plays a crucial role in scaling AI. By leveraging open-source technologies, enterprises can benefit from a collaborative approach to AI development, accessing a wealth of community-driven innovations and best practices. “The open-source community is a vital component of AI advancement,” Gil highlighted. “It fosters a spirit of collaboration and continuous improvement, essential for scaling AI effectively across enterprises.”
As enterprises navigate the complexities of scaling AI, IBM’s comprehensive approach—spanning advanced technologies, robust data management, continuous learning, and open-source collaboration—provides a clear pathway to success. “Scaling AI is a journey,” Gil concluded. “It’s about creating a sustainable, adaptable framework that grows with the enterprise, driving innovation and competitive advantage at every step.”
As IBM continues to push the boundaries of AI, the future holds immense potential for enterprises willing to embrace these transformative technologies. Darío Gil’s vision for AI is one where innovation and collaboration drive progress, ensuring that AI serves not just as a tool for efficiency but as a catalyst for groundbreaking advancements across industries.
One of the key areas of focus for IBM moving forward is the integration of AI with other cutting-edge technologies, such as quantum computing and blockchain. “The convergence of AI with quantum computing can unlock new levels of problem-solving capabilities that were previously unimaginable,” Gil noted. “By combining the strengths of these technologies, we can tackle some of the most complex challenges facing humanity, from climate change to healthcare.”
IBM is also committed to ensuring that AI development remains ethical and inclusive. The company is actively working on initiatives to address biases in AI models and to promote transparency and accountability in AI systems. “As we look ahead, it’s crucial that we build AI that is fair, transparent, and respects the values of our society,” Gil emphasized. “We are dedicated to leading the charge in creating ethical AI frameworks that benefit everyone.”
In enterprise applications, IBM plans to expand its portfolio of AI-driven solutions, providing businesses with even more tools to enhance their operations and drive innovation. The company’s continued investment in research and development ensures its clients have access to the latest advancements in AI technology. “Our goal is to empower enterprises to leverage AI in ways that were previously thought impossible,” Gil said. “We are constantly exploring new frontiers and developing solutions that will keep our clients at the forefront of their industries.”
Moreover, IBM’s commitment to open-source AI models will play a significant role in the future of AI development. By fostering a collaborative environment, IBM aims to accelerate the pace of innovation and ensure that AI technology evolves in a way that is beneficial for all stakeholders. “The future of AI is one that is built on collaboration and shared knowledge,” Gil stated. “By embracing open-source principles, we can create a thriving ecosystem where everyone has the opportunity to contribute and benefit from AI advancements.”
As the landscape of AI continues to evolve, IBM remains steadfast in its mission to drive technological progress while addressing its ethical and societal implications. “The road ahead is full of exciting possibilities,” Gil concluded. We are committed to leading the way in AI innovation, ensuring that our advancements serve the greater good and pave the way for a better future for all.”
With a forward-looking approach that combines technological excellence, ethical considerations, and a collaborative spirit, IBM is well-positioned to shape the future of AI and drive meaningful change across the globe. As enterprises prepare to navigate this dynamic landscape, they can look to IBM for guidance, support, and innovative solutions to help them thrive in the age of AI.
]]>A Three-Day Developer Extravaganza
The evolution of Google I/O over the past few years has been notable. The event was canceled in 2020, and the 2021 edition was a modest affair, streamed to a limited live audience in Mountain View. In 2022 and 2023, attendees were invited for just one day. This year, however, marked a significant shift. After the keynote, Google hosted live sessions for in-person attendees and organized after-hour social events, creating a vibrant atmosphere of collaboration and learning.
Exclusive Programming for Googlers
While the pre-recorded live sessions released on YouTube gave the impression of a three-day event, there was another day of programming exclusively for Googlers at the Shoreline Amphitheatre. CEO Sundar Pichai, who hosted the event, revealed in an internal email that thousands of Googlers attended, with many more streaming it internally. Pichai shared images on LinkedIn, capturing the excitement and engagement of the event.
Android engineering VP Dave Burke and teams from Google DeepMind, Search, and Labs demonstrated the innovations announced earlier in the week. Project Astra, a conference highlight, was showcased again, with some announcements made available to employees for internal testing.
Announcing the Gemini Hackathon
The highlight of the Demo Slam was Pichai’s announcement of an internal hackathon encouraging Google employees to experiment with Gemini, Google’s AI project. This initiative aims to foster AI experimentation and could potentially lead to new product developments. Googlers are encouraged to form teams and collaborate, with Google executives selecting finalists to present at a company-wide meeting. The hackathon also offers a monetary prize for the winning teams.
Pichai emphasized the importance of this initiative, stating, “We want to create more opportunities for us to come together as a company in the spirit of innovation and problem-solving, focused on our biggest opportunities like AI.”
A Spirit of Innovation
The Gemini hackathon is designed to ignite a spirit of innovation and problem-solving among Google’s workforce. By encouraging employees to collaborate and experiment, Google aims to leverage its internal talent to push the boundaries of AI technology. This hackathon reflects Google’s broader strategy to integrate AI into its products and services, ensuring the company remains at the forefront of technological innovation.
As Pichai noted, “Our goal is to harness the collective creativity and expertise of our employees to drive the next wave of AI advancements. The Gemini hackathon is a key step in that direction, fostering a culture of innovation and collaboration.”
Google I/O 2024 has set a new standard for developer conferences, blending public engagement with exclusive internal initiatives. The introduction of the Gemini hackathon underscores Google’s commitment to AI and its belief in the power of its employees to shape the future of technology. As the hackathon progresses, the tech world will be watching to see what groundbreaking innovations emerge from this exciting initiative.
]]>AI is a controversial topic within the Linux community, with some using the open source OS specifically to avoid using things like AI. Despite the controversy, Red Hat appears to be throwing its support behind the burgeoning tech, rolling out a version of RHEL specifically designed “to seamlessly develop, test and run best-of-breed, open source Granite generative AI models to power enterprise applications.”
The main objective of RHEL AI and the InstructLab project is to empower domain experts to contribute directly to Large Language Models with knowledge and skills. This allows domain experts to more efficiently build AI-infused applications (such as chatbots).
Red Hat hopes to challenge the status quo, in which many of the Large Language Models (LLMs) are based on heavily patented, closed source licenses. In addition, training LLMs can be expensive and often does not prioritize privacy, confidentiality, or data sovereignty.
Red Hat (together with IBM and the open source community) proposes to change that. We propose to introduce the familiar open source contributor workflow and associated concepts like permissive licensing (e.g. Apache2) to models and the tools for open collaboration that enable a community of users to create and add contributions to LLMs. This will also empower an ecosystem of partners to deliver offerings and value to enable extensions and the incorporation of protected information by enterprises.
Those interested in Red Hat’s approach to AI can learn more here and join the open source community and start contributing here.
]]>The first test focused on text summarization, a critical function for many AI applications. The models were tasked with summarizing a lengthy article into both a short, 2-3 sentence summary and a more detailed 5-6 sentence version. GPT-4o delivered summaries that were not only concise but also clear and well-structured. In contrast, GPT-4, while accurate, tended to adopt a more promotional tone, which was less suited for a neutral summary.
“GPT-4o’s tone was impressive,” remarked Adibs. “It managed to capture the essence of the content without sounding like an advertisement, which is exactly what we were looking for.”
In practical terms, GPT-4o can provide users with concise and informative summaries suitable for various applications, from academic research to business reports. The ability to distill complex information into clear, concise summaries can significantly enhance productivity and comprehension, making it easier for users to quickly digest large volumes of information.
Moreover, GPT-4o’s enhanced summarization capabilities suggest improvements in areas such as customer service, where quick and accurate information retrieval is essential. By providing more precise and contextually appropriate summaries, GPT-4o can help businesses improve their response times and overall customer satisfaction.
Creating a compelling product description is essential for capturing the attention of potential customers. In this test, GPT-4o was tasked with writing a concise, punchy product description for a hypothetical software tool that tracks social media analytics. The model’s output was impressive, delivering a dynamic and engaging description that effectively highlighted the key benefits of the software. This capability is crucial for marketers who must craft messages that resonate quickly and powerfully with their audience.
GPT-4o’s ability to generate high-quality marketing content demonstrates its advanced natural language processing skills. It not only understands the product’s core features but also conveys its value proposition in an attractive and persuasive way. Adibs commented, “The precision and flair with which GPT-4o crafts product descriptions show its potential to revolutionize marketing communications. Businesses can leverage this to create impactful, concise content that drives engagement and conversions.”
Moreover, the practical implications of GPT-4o’s prowess in marketing extend beyond mere product descriptions. This model can be utilized for various marketing materials, including social media posts, email campaigns, and even advertisement copy. Its ability to maintain a consistent tone and deliver clear, compelling messages can help businesses streamline their marketing efforts, ensuring that their communications are both effective and efficient. By automating these aspects, companies can focus more on strategy and creativity, leveraging AI to handle the repetitive and time-consuming task of content creation.
Comprehending and analyzing visual data is a significant leap forward in AI capabilities. GPT-4o’s performance in the multimodal understanding test showcased its prowess in this area. When tasked with analyzing an image and explaining it in a table format, GPT-4o demonstrated a remarkable ability to interpret and structure visual information accurately. This feature is especially beneficial for applications requiring textual and visual data integration, such as medical imaging, autonomous vehicles, and advanced data analysis.
GPT-4o’s multimodal capabilities extend beyond simple image recognition. It can understand complex visual contexts and generate detailed, structured outputs that are easy to interpret. Saj Adibs, CEO of Skill Leap AI, highlighted the importance of this feature: “The integration of visual and textual data is crucial for developing more sophisticated AI applications. GPT-4o’s ability to seamlessly handle both types of information opens up new possibilities for innovation in various industries.”
These advancements in multimodal understanding enhance the AI’s utility in professional settings and make it more accessible for everyday use. For instance, users can employ GPT-4o to analyze charts, graphs, and other visual aids, providing clear, concise summaries that enhance comprehension and decision-making. This capability is particularly valuable in educational tools, where AI can assist students in understanding complex visual materials, enhancing their learning experience. As AI continues to evolve, integrating multimodal understanding will play a pivotal role in expanding its applications and improving its effectiveness.
Image generation has become a hallmark of advanced AI capabilities, and GPT-4o demonstrates significant advancements in this realm. In the head-to-head comparison, GPT-4o produced a compelling and visually detailed image of two AI robots in battle. The level of detail, composition, and creativity surpassed that of GPT-4, showcasing the new model’s ability to generate high-quality visual content from textual prompts. This enhancement is a testament to the model’s improved algorithms and its ability to interpret and visualize complex ideas creatively.
Saj Adibs, CEO of Skill Leap AI, emphasized the potential applications of this feature: “The advancements in image generation by GPT-4o open up new frontiers for creative industries. From digital art to marketing campaigns, the ability to generate high-quality, customized images on demand will revolutionize how businesses approach visual content creation.” The ability to quickly generate detailed and aesthetically pleasing images can save businesses significant time and resources, allowing them to focus more on strategy and less on production.
Moreover, the implications for education and training are profound. Teachers and trainers can use GPT-4o to create illustrative content that enhances learning experiences, making complex subjects more accessible and engaging. For example, medical students can benefit from AI-generated diagrams and simulations, while history students might explore detailed reconstructions of historical events. As AI image generation continues to evolve, it promises to unleash unprecedented levels of creativity and efficiency across various fields, solidifying its role as an indispensable tool in the digital age.
The research capabilities of AI models have always been a significant point of comparison, and GPT-4o stands out with its enhanced performance. In the head-to-head test, GPT-4o demonstrated a remarkable ability to conduct in-depth research, identifying specific use cases, potential benefits, and challenges of AI in the accounting industry. The model provided comprehensive information and included relevant links to articles and reports, ensuring users could delve deeper into the topics if needed.
Saj Adibs, CEO of Skill Leap AI, highlighted the importance of this feature: “GPT-4o’s research capabilities mark a significant advancement in AI technology. The ability to quickly gather, synthesize, and present detailed information from various sources is invaluable for professionals across industries. It transforms how we approach problem-solving and decision-making.” This feature is particularly beneficial for fields that require extensive research, such as academia, legal, and medical professions, where accuracy and depth of information are crucial.
Additionally, GPT-4o’s ability to provide contextually relevant and well-structured information enhances its utility for everyday users. Whether it’s students conducting research for their assignments or business professionals preparing reports, the model’s ability to deliver precise and detailed information expedites the research process and ensures high-quality outputs. As AI continues to evolve, these research capabilities will likely become even more refined, further solidifying the role of AI as a critical tool for knowledge acquisition and dissemination.
One of the most practical applications of AI models is their ability to generate and debug code. In comparing GPT-4 and GPT-4o, the latter demonstrated a clear edge in this domain. When tasked with generating Python code for a simple snake game, GPT-4o produced functional code and offered a more interactive and user-friendly version of the game. The code included features such as dynamic speed adjustments and a scoring system, making the game more engaging for users.
Saj Adibs, CEO of Skill Leap AI, emphasized the significance of this capability: “The ability of GPT-4o to generate high-quality, functional code is a game-changer for developers. It accelerates the development process and ensures that the code is robust and optimized.” This capability is particularly beneficial for developers working under tight deadlines or needing to quickly prototype and test new ideas. The practical application of AI in code generation extends beyond simple tasks, with potential use cases in complex software development, debugging, and even automating routine coding tasks.
Moreover, GPT-4o’s step-by-step guidance on running the generated code is invaluable for beginners who may not have extensive programming knowledge. This feature lowers the barrier to entry for learning programming, making it accessible to a broader audience. By simplifying the process and providing clear instructions, GPT-4o empowers users to confidently tackle more complex projects. As AI continues to integrate into the software development lifecycle, its ability to generate and refine code will undoubtedly transform the industry, making it more efficient and innovative.
The head-to-head comparison between GPT-4 and GPT-4o reveals that the latter has made significant strides in various AI functionalities, from text summarization and product descriptions to multimodal understanding and code generation. GPT-4o consistently demonstrated superior performance in these tasks, showcasing its enhanced capabilities and practical applications. The advancements in GPT-4o highlight OpenAI’s commitment to pushing the boundaries of what AI can achieve, offering users a more robust and versatile tool.
Saj Adibs, CEO of Skill Leap AI, summed up the implications of these advancements: “GPT-4o represents a pivotal moment in the evolution of AI. Its ability to perform complex tasks with greater accuracy and efficiency is a testament to our rapid progress in this field. This model not only meets the demands of today’s users but also paves the way for future innovations.” The improved performance of GPT-4o suggests that AI will continue to play an increasingly integral role in various industries, driving innovation and efficiency.
However, as impressive as GPT-4o is, it also raises questions about the future of AI development and the potential challenges that come with it. Ethical considerations, data privacy, and the impact on employment must be carefully navigated. The AI community must address these concerns to ensure that the benefits of AI advancements are maximized while minimizing potential risks. As we look ahead, GPT-4o stands as a beacon of what is possible, but it also serves as a reminder of the responsibilities of such powerful technology.
Adibs concluded, “The future of AI looks promising with advancements like GPT-4o. It’s an exciting time for the industry, and we look forward to seeing how these tools will continue to evolve and impact our world.”
]]>Mira Murati, OpenAI’s Chief Technology Officer, led the presentation, highlighting the importance of this new development. “Our mission is to democratize AI, ensuring that everyone, regardless of their economic status, has access to our most advanced models. The introduction of the desktop app is a monumental step in that direction,” Murati said.
The new ChatGPT desktop app is designed to provide a seamless and intuitive user experience. It can be opened quickly with a keyboard shortcut, allowing users to ask ChatGPT questions without disrupting their workflow. The app opens in a window on the screen and can interact with users based on what’s displayed, offering capabilities like screenshot analysis and contextual discussions. This level of integration makes it a powerful tool for professionals who rely on AI assistance for coding, research, and other tasks.
Murati emphasized, “We have overhauled the user interface to make the experience more intuitive and seamless, allowing users to focus on collaboration rather than navigating complex interfaces. This new app aims to enhance productivity and efficiency, particularly for those who use ChatGPT for complex tasks like coding and data analysis.”
The new desktop app’s initial release on macOS highlights OpenAI’s commitment to providing high-quality, platform-specific experiences. Mac users will benefit from unique features that leverage the macOS environment, including optimized performance and integration with Mac-specific functionalities. “We’re excited to offer Mac users a first-class experience with the ChatGPT desktop app, taking full advantage of the macOS ecosystem,” Murati said.
The app was announced in tandem with the launch of GPT-4o, OpenAI’s latest model, which integrates text, speech, and vision capabilities. Once GPT-4o is fully rolled out, users will be able to use Voice Mode in the desktop app to have conversations with ChatGPT. This feature was demonstrated during the livestream using coding examples, showcasing the model’s ability to handle real-time conversational speech, interruptions, and contextual shifts. “The integration of multimodal functions makes GPT-4o a versatile tool that can adapt to a wide range of scenarios and needs,” Murati noted.
The new desktop app marks a significant improvement in accessibility for ChatGPT. Previously, users could only access ChatGPT through third-party apps and browser extensions. With this first-party app, OpenAI ensures a more reliable and integrated experience. Both free and paid users will have access to the app, which is available starting today for Plus users and will roll out to free users in the coming weeks. OpenAI also announced plans to launch a Windows version of the app later this year, although specific dates have not been provided.
Sam Altman, CEO of OpenAI, commented, “We are thrilled to bring the power of ChatGPT to the desktop. This app will make it easier for users to integrate AI into their daily routines, whether they’re coding, conducting research, or just seeking information.”
The introduction of the ChatGPT desktop app is poised to have a significant impact on both professional and personal use cases. For professionals, particularly those in tech and creative industries, the app provides an invaluable tool for enhancing productivity and innovation. Developers, for example, can use the app to get real-time coding assistance, troubleshoot errors, and receive detailed explanations of complex concepts. The ability to interact with ChatGPT through voice commands further enhances the convenience and utility of the app.
In personal use, the app’s capabilities extend to a variety of tasks, from learning new languages to getting help with homework. The ease of access and intuitive interface make it a useful tool for students, educators, and hobbyists alike. “Our goal is to make AI accessible to all, enabling everyone to benefit from its potential,” Murati emphasized.
Looking ahead, OpenAI is committed to continuously improving the ChatGPT desktop app and expanding its capabilities. Future updates are expected to enhance the integration of multimodal functions, making the app even more powerful and versatile. OpenAI’s ongoing research and development efforts will ensure that the app stays at the forefront of AI technology, providing users with the most advanced tools available.
Murati highlighted OpenAI’s dedication to ethical AI development: “Safety and security will remain top priorities. We collaborate with various stakeholders, including academic institutions, policymakers, and civil society organizations, to develop robust safety protocols and ensure responsible use of our technology. Ethics and responsibility are at the core of our mission.”
The launch of the first-party ChatGPT desktop app represents a significant milestone in OpenAI’s mission to democratize AI. By providing a seamless, integrated experience that leverages the advanced capabilities of GPT-4o, OpenAI is setting a new standard for AI interaction. As the app becomes more widely adopted, its influence will undoubtedly grow, shaping the future of AI and its role in society. With OpenAI at the helm, the potential for AI to drive positive change and innovation is immense.
In summary, the new ChatGPT desktop app is a game-changer, offering enhanced accessibility, multimodal capabilities, and a user-friendly interface. It reflects OpenAI’s vision for a more inclusive and powerful AI future, ensuring that advanced technology is within reach for everyone. As the technology continues to evolve, the ChatGPT desktop app is set to become an indispensable tool for both personal and professional use, transforming how we work, learn, and interact with AI.
]]>As Big Tech companies reported their latest earnings, they also unveiled a massive wave of planned spending on AI projects and infrastructure. Alphabet, Meta, and Microsoft are leading the charge, which have pulled ahead in the race, while Apple faces mounting challenges. Dan Ives, Managing Director at Wedbush Securities, noted, “The fourth industrial revolution has begun.” According to Ives, Apple’s forthcoming Worldwide Developers Conference (WWDC) will be crucial, calling it “probably the most important event for Apple that we’ve seen potentially in a decade.”
Ives is confident that Apple’s AI strategy will be laid out clearly at WWDC. He anticipates that CEO Tim Cook will reveal a new AI app store and proprietary AI technology for the iPhone 16. “This is the start of AI coming to Cupertino,” he added.
While some tech companies may face challenges, others like Salesforce, Oracle, Palantir, and SoundHound AI have shown robust numbers and benefit from this tidal wave. “If you look at these hyperscale players like Microsoft, Amazon, and Google, their spending on AI is just beginning,” Ives said. He also emphasized the broad impact of the AI revolution, noting, “You have to own the right plays in semis, software, and infrastructure.”
Apple’s Strategy and Challenges
At the forefront of Apple’s AI ambitions is the Worldwide Developers Conference, where Cook is expected to lay out the company’s strategy for AI. “It’s probably the most important event for Apple in a decade,” said Ives, emphasizing the significance of AI on the services side for developers. He believes this could lead to the creation of an “AI app store.”
Despite the potential, Apple has faced criticism for being behind in the AI race. However, Ives remains optimistic about the company’s trajectory. “Behind the scenes, were they behind when it came to AI? Yeah, but I think they’ve quickly caught up,” he noted. He highlights Apple’s strong installed base of 2.2 billion devices worldwide and calls this “the start of AI coming to Cupertino.”
Ives also points to Apple’s apology for its ad campaign as evidence of its commitment to quality. “When you make the best products in the world, you have a spotlight on you,” he said, predicting that the iPhone 16 will help Apple return to growth.
Fourth Industrial Revolution and Market Dynamics
Ives considers the current AI revolution to be the fourth industrial revolution, already reshaping markets. “We’re not even in the first inning; I’d say we’re in the dugout,” he said, underscoring that companies like Dell and Oracle, which weren’t traditionally seen as AI players, are now getting significant traction.
Despite concerns about potential over-investment in AI, Ives suggests a diversified approach. “There’s almost a basket approach to this,” he said. However, he cautions against over-optimism, advising investors to focus on companies like Microsoft, Amazon, and Google.
Ives also touched upon the emerging competitive landscape between U.S. and Chinese companies, noting the Biden Administration’s tariffs and the potential retaliation from China. “Chinese EVs are going to come here at one point and will be successful,” he predicted.
Conclusion: Apple’s Bright Future Amid Uncertainties
The AI revolution is gaining momentum, and tech giants like Apple and Microsoft are at the forefront of this transformation. While Apple may have initially lagged behind in the AI race, Ives believes the company is poised for a comeback, calling it a “Renaissance of growth.”
He is confident that Apple’s AI strategy will yield significant returns, particularly with the launch of the iPhone 16. “This is an iPhone 16 AI super cycle,” he emphasized, predicting that investors will view the present moment as a golden buying opportunity.
Ultimately, Ives advises investors to focus on the right AI plays across semiconductors, software, and infrastructure. With Apple’s WWDC and its strategic focus on AI, the tech giant is positioning itself for success in the fourth industrial revolution.
Apple’s Worldwide Developers Conference (WWDC) is widely anticipated to be a defining moment for the company’s AI ambitions. Dan Ives sees the event as pivotal in solidifying Apple’s role in the rapidly evolving AI landscape. He emphasized, “It’s probably the most important event for Apple in a decade.” The unveiling of a new AI-focused strategy will be crucial for reassuring investors and developers alike.
Tim Cook is expected to introduce a strategic framework with a dedicated AI app store and proprietary AI technology integrated into the iPhone 16, set to launch in September. The move marks a significant step toward integrating AI into Apple’s ecosystem. “Cook laying out the AI strategy on the services side for developers is going to be the start of an AI app store,” Ives noted.
Apple AI App Store for Developers
This AI app store will offer developers the tools to build and distribute AI-powered applications, thus expanding Apple’s footprint in this emerging market. Additionally, the company is expected to showcase enhancements to Siri and other AI-driven features that will distinguish its services from competitors.
Despite lagging behind industry leaders like Microsoft and Google, Ives believes Apple has a distinct advantage due to its loyal customer base and extensive hardware ecosystem. “They’ve quickly caught up, and you have the best-installed base of 2.2 billion in the world,” he said. With a strong foundation, Apple aims to leverage AI to enhance its existing products and services.
Balance Between Innovation and Privacy
Moreover, Apple’s AI strategy is expected to address privacy concerns that have plagued the industry. The company is renowned for its emphasis on user privacy, and this will likely be a key differentiator in how it positions its AI offerings. By integrating AI within its tightly controlled ecosystem, Apple can ensure a balance between innovation and privacy.
In light of these developments, Ives remains optimistic about Apple’s future growth. He predicts the company will recover from recent challenges and enter a “Renaissance of growth” driven by the iPhone 16 and AI. “This iPhone 16 AI super cycle will play out,” he affirmed. Despite investor skepticism, he believes the WWDC will begin a new chapter for Apple in the AI race.
The strategic focus on AI will significantly impact Apple’s bottom line. Dan Ives anticipates that the forthcoming iPhone 16 and the expansion of AI services will drive a new growth cycle for the tech giant. “I think it’s going to be the last quarter of negative iPhone growth,” Ives remarked, emphasizing that this will herald a “Renaissance of growth” for the company.
Apple’s ability to integrate AI seamlessly into its hardware and services is at the core of this optimistic outlook. The proprietary AI technology expected to be embedded in the iPhone 16 will likely enhance the user experience across the board, making the new model particularly attractive to consumers. Improvements to Siri, personalized recommendations, and new camera features are among the anticipated AI-driven enhancements.
AI App Store To Open Up New Revenue Streams
Moreover, the anticipated AI app store will open up new revenue streams, allowing Apple to monetize developer innovation through a diverse ecosystem of AI applications. The strategy could mirror the success of Apple’s existing app store, which has consistently been a significant contributor to its services revenue. This shift will accelerate Apple’s services growth, including Apple Music, iCloud, and Apple TV+.
In addition to the consumer-focused AI strategy, Apple is poised to bolster enterprise offerings. Integrating AI into productivity tools like Pages, Numbers, and Keynote could challenge Microsoft and Google in the business sector. Ives pointed out that Apple’s significant market presence gives it a unique advantage in expanding its AI services. “Apple is where Meta was 18 months ago,” he stated, predicting a similar trajectory of rapid AI growth.
AI Apps To Drive the ‘iPhone Growth Story’
Despite recent headwinds in the global smartphone market, Apple’s diversified product portfolio and its strategic investments in AI positions the company well for long-term success. Ives noted that investors may be underestimating Apple’s potential. “I don’t think numbers are reflecting what ultimately is going to happen on the services side and the iPhone growth story,” he said.
Apple’s renewed focus on AI will also help it differentiate itself in an increasingly competitive landscape. By leveraging its unparalleled ecosystem, the company can create a tightly integrated, seamless experience that competitors may find difficult to replicate. If successful, this approach will solidify Apple’s status as a leader in the fourth industrial revolution and significantly boost its bottom line.
The rapid advancement of AI technology has created a seismic shift across multiple industries, compelling tech giants like Apple, Microsoft, and Alphabet to intensify their investments. For Dan Ives, this wave of innovation represents a pivotal moment in technological history, akin to the early days of the Internet revolution. He said, “This is a 1995 moment, not a 1999 moment.” He sees this period as an opportunity for investors to capitalize on companies leading the AI transformation.
Apple’s AI strategy, despite initially being perceived as lagging behind, is now positioned to play a significant role in the company’s growth trajectory. With its proprietary AI technology set to debut in the iPhone 16 and the anticipated launch of an AI-focused App Store, Apple aims to capture a new market segment that could drive significant revenue growth. Ives described this upcoming phase as “probably the most important event for Apple that we’ve seen potentially in a decade.”
The AI Revolution Offers a Rare Opportunity
The broader implications of AI’s integration into the tech industry are profound. Companies like Microsoft, Alphabet, and Meta already leverage AI to transform their core businesses. This shift leads to what Ives calls “the fourth industrial revolution,” with AI fundamentally changing how products are built, services are delivered, and businesses are managed.
For investors, the AI revolution offers a rare opportunity to capitalize on transformative technologies that will shape the future. Whether it’s Apple’s foray into AI-powered iPhones, Microsoft’s bold investment in OpenAI, or Alphabet’s continued dominance in search and advertising, the potential rewards are immense. However, as with any revolutionary period, the risks are equally significant, requiring careful navigation of the competitive landscape and global economic challenges.
Ultimately, as the tech titans battle for dominance in the AI era, investors are betting on AI’s future as a transformative force that will redefine industries, create new markets, and unlock unprecedented economic opportunities. Ives remains confident that those who align their investments with this wave of innovation will emerge as the true winners of the fourth industrial revolution.
]]>Microsoft’s recent announcement to invest $3 billion in Wisconsin marks a pivotal moment in the AI landscape. By partnering with the Green Bay Packers, the tech giant aims to establish a cutting-edge hub for startups focusing on manufacturing and artificial intelligence research. This strategic move aligns with Microsoft’s broader vision of fostering regional innovation and underscores the emerging trend of vertical integration in AI.
Steve Case, the co-founder of AOL and current chairman and CEO of Revolution, views Microsoft’s investment as a significant shift from broad, horizontal AI platforms to industry-specific vertical applications. “We’re seeing a transition from AI being really about the big horizontal platforms… to now more vertical AI, which creates an opportunity all around the country,” Case said in an interview with CNBC.
What Is Vertical AI?
Vertical AI refers to artificial intelligence applications tailored for specific industries, such as manufacturing, healthcare, and agriculture, unlike horizontal AI platforms like OpenAI’s GPT-4 or Google’s Bard, which offer broad-based capabilities across sectors, vertical AI hones in on specialized problems and delivers customized solutions. This investment aims to position Wisconsin as a leader in this new wave of AI innovation.
Creating a Hub for Manufacturing Innovation
The investment will establish a hub where startups can collaborate and research new ways to integrate AI into traditional manufacturing processes. This approach emphasizes how technology can transform existing industries rather than disrupt them. With Microsoft’s commitment, the Wisconsin hub is poised to become a model for other regions seeking to foster innovation outside Silicon Valley.
Breaking Down the Partnership
Steve Case’s Perspective
Steve Case sees Microsoft’s investment aligning with his vision of distributing tech opportunities across America. “We can’t just have AI be big tech getting bigger; we can’t just have AI being Silicon Valley continuing to be dominant,” he said. For Case, vertical AI represents the future of innovation, where startups nationwide can tap into AI’s transformative potential to reimagine industries and drive productivity.
By emphasizing vertical integration and regional innovation, Microsoft’s investment promises to create new opportunities for startups, entrepreneurs, and established businesses. As Wisconsin becomes a hub for manufacturing-focused AI, this initiative will likely serve as a blueprint for how AI can transform local economies and drive nationwide growth.
Artificial intelligence is entering a new phase of innovation, marked by the shift from general-purpose platforms to specialized, industry-focused applications. This emerging trend, known as vertical integration in AI, is poised to transform the technology landscape across diverse sectors, from agriculture to healthcare.
What Is Vertical Integration in AI?
Vertical integration in AI refers to creating industry-specific applications that address unique challenges and deliver tailored solutions. Unlike horizontal AI platforms, such as OpenAI’s GPT-4 and Google’s Bard, which offer broad, cross-industry functionality, vertical AI targets specialized problems and optimizes business processes within particular domains.
Examples of Vertical AI
Steve Case’s Perspective on Vertical AI
In a recent CNBC interview, Steve Case emphasized the importance of vertical AI for regional innovation. “We’re seeing a transition from AI being really about the big horizontal platforms… to now more vertical AI, which creates an opportunity all around the country,” he said. For Case, vertical integration represents an opportunity to distribute technological benefits beyond Silicon Valley.
Why Is Vertical AI Important?
Microsoft’s Investment in Vertical AI
Microsoft’s $3 billion investment in Wisconsin exemplifies this shift toward vertical integration in AI. The partnership between Microsoft and the Green Bay Packers aims to establish a manufacturing-focused hub where startups can develop AI applications tailored for the industry. This initiative promises to transform Wisconsin into a leader in AI-driven manufacturing innovation.
Impact on the Future of Work
Vertical AI’s rise will also have profound implications for the future of work. As AI becomes more embedded in specific industries, there will be a growing need for professionals who understand AI integration’s business and technical aspects. Skills like creativity, collaboration, and critical thinking will become increasingly important, complementing the traditional emphasis on coding.
The rise of vertical integration in AI marks a new chapter in the technology landscape, offering opportunities for regional innovation, productivity gains, and industry transformation. By focusing on specialized applications, vertical AI has the potential to redefine how businesses operate, unlocking new levels of efficiency and driving growth in ways previously unimaginable.
As artificial intelligence (AI) advances, the future of work is transforming, redefining job roles, required skills, and productivity across various industries. This evolution is technological and cultural, affecting how and where people work.
Disruption and Opportunity in Job Roles
AI’s growing presence in the workplace brings both disruption and opportunity. While some routine tasks become automated, new roles emerge, often demanding higher-level skills.
According to Steve Case, co-founder of AOL, “Not everybody should focus on coding. Some other skills around communication, collaboration, and creativity will be more important.” His words reflect the broader shift toward roles that blend technical expertise with soft skills.
Essential Skills in the AI Era
In the era of AI, success requires a combination of technical and non-technical skills.
The Hybrid Work Model
The pandemic has accelerated the adoption of remote work, leading to hybrid models combining in-office and remote work.
Many startups embrace the hybrid model, offering flexibility while benefiting from in-person collaboration. Steve Case noted, “We’re moving from a world where everybody was in the office five years ago to now more of a hybrid concept.”
Productivity Gains and Regional Dispersion
AI promises significant productivity gains across sectors, potentially reducing the need for some jobs while creating others.
Future Workforce Development
To harness the full potential of AI, the workforce will need reskilling and upskilling programs.
Both challenges and opportunities mark the evolution of work in the era of AI. While some job roles will be disrupted, new opportunities will emerge, particularly for those willing to adapt and acquire new skills. Embracing this shift will require a concerted effort from governments, companies, and educational institutions to ensure the workforce is prepared for an AI-driven future.
As the work landscape changes dramatically with the rise of artificial intelligence (AI), preparing for the future has become a crucial priority for businesses, governments, and individuals. This preparation involves addressing challenges, building new skills, and creating inclusive opportunities to ensure the future workforce thrives.
Understanding the Changing Skill Set
One of the critical aspects of preparing for the future of work is recognizing the shifting skill requirements.
Steve Case emphasizes the importance of blending technical and soft skills, saying, “Some other skills around communication, collaboration, and creativity are going to be more important.”
Government and Policy Initiatives
Governments play a vital role in shaping the future workforce through policy initiatives and strategic investments.
Corporate Strategies and Workforce Development
Businesses also play a pivotal role in preparing their employees for the future.
Educational Reform and Lifelong Learning
Educational institutions need to update their curricula to reflect the changing demands of the workforce.
Inclusivity and Equity in the Future Workforce
Ensuring that the benefits of AI are distributed equitably is crucial for a sustainable future.
Conclusion
Preparing for the future of work requires a collaborative approach that involves governments, businesses, educational institutions, and individuals. By understanding the changing skill set, investing in strategic initiatives, fostering a culture of continuous learning, and promoting inclusivity, we can build a workforce that thrives in the era of AI. The key lies in embracing change, acquiring new skills, and creating opportunities for all to participate in this transformative journey.
In the rapidly evolving landscape of artificial intelligence and technological advancements, fostering a more inclusive innovation economy is crucial. As the U.S. and global economies transition toward AI and digital transformation, ensuring broad participation and equitable growth becomes imperative. Steve Case, co-founder of AOL and Revolution Chairman, emphasizes the importance of building a future where more people and regions have opportunities to thrive.
Regional Dispersion of Innovation
One significant shift in the innovation economy is the regional dispersion of tech talent and investment. Traditionally, Silicon Valley, New York, and other coastal cities have been dominant hubs for tech innovation, but a new trend is emerging.
“Fostering a more inclusive innovation economy involves investing in tech hubs so that it’s not just a few places doing well,” Case emphasizes.
Diverse and Inclusive Workforce Development
Ensuring a diverse and inclusive workforce is critical to building an equitable innovation economy. Key strategies include:
Accessible Education and Lifelong Learning
Inclusive innovation starts with accessible education and continuous learning opportunities.
Equitable Funding and Support for Startups
Access to funding and support remains a challenge for many startups, particularly those led by women, minorities, and those located outside traditional tech hubs.
Collaboration Between Public and Private Sectors
Building an inclusive innovation economy requires collaboration between various stakeholders.
Conclusion
Fostering a more inclusive innovation economy is not just a matter of social equity but a strategic imperative for sustainable growth. By promoting regional dispersion of innovation, ensuring diverse workforce participation, providing accessible education, and supporting underrepresented startups, we can build a future where technology drives opportunity for all. The vision is clear: an innovation economy that truly reflects the diversity of talent and ambition across America and beyond.
As the world stands on the cusp of an unprecedented technological revolution driven by artificial intelligence, investing in the future has never been more critical. This involves financial commitments, strategic foresight, inclusive policies, and comprehensive support systems that empower all regions and people to thrive in the emerging digital economy.
Microsoft’s $3 billion investment in Wisconsin clearly shows how strategic investments can transform regional economies and foster innovation. The company is setting the stage for Wisconsin to become a hub of technological advancement by establishing a new AI facility. This initiative underscores the transition from horizontal AI platforms toward a more vertical integration approach, where specific industries like manufacturing and agriculture can leverage AI to drive productivity and growth.
Steve Case’s advocacy for inclusive innovation and regional dispersion is a powerful reminder that the future of technology should not be concentrated in a few dominant tech hubs. Instead, the benefits of AI and digital transformation should be shared across regions and communities, giving rise to a more equitable and inclusive economy. His work through the “Rise of the Rest” initiative demonstrates how targeted investments and support can revitalize local ecosystems, helping startups scale and compete globally.
The evolution of work in the AI era will present challenges and opportunities. As automation and machine learning redefine job roles, there will be a growing need for new skills and adaptability. Companies and governments must prioritize workforce development through reskilling and upskilling programs to prepare workers for the jobs of the future. This includes fostering creativity, critical thinking, and collaboration—uniquely human skills that will remain in demand.
Moreover, fostering a more inclusive innovation economy requires breaking down barriers that prevent underrepresented groups and regions from fully participating in the tech industry. This means closing the gender and racial gaps in tech education and employment, providing equitable funding for startups, and creating supportive workplace cultures that value diverse perspectives.
The future is uncertain, but the path forward is clear. Investing in regional innovation, inclusive policies, workforce development, and education will build a resilient and prosperous economy. By working together—businesses, governments, and communities—we can ensure that the AI revolution benefits all, creating a future where opportunity is accessible to everyone.
]]>