Your browser does not support JavaScript!

Comparative Analysis of AI Chatbot Subscriptions: ChatGPT, Gemini, and Meta AI

GOOVER DAILY REPORT August 12, 2024
goover

TABLE OF CONTENTS

  1. Summary
  2. Subscription Models and Features
  3. Performance Comparison: Google Gemini vs OpenAI ChatGPT
  4. Performance Comparison: Meta AI vs ChatGPT vs Google Gemini
  5. Key Findings and Practical Implications
  6. Conclusion

1. Summary

  • The report titled 'Comparative Analysis of AI Chatbot Subscriptions: ChatGPT, Gemini, and Meta AI' offers an in-depth comparison of three prominent AI chatbots: OpenAI's ChatGPT, Google's Gemini, and Meta's AI. It examines their subscription models, performance across various tasks, and unique strengths and weaknesses. Key findings highlight their utility in applications ranging from coding and creative writing to ethical decision-making. ChatGPT Plus offers advanced features like GPT-4 and DALL-E 3 integration but lacks storage options found in Google's Gemini Advanced, which bundles additional cloud storage. The performance analysis shows Google Gemini excelling in creative tasks and coding, while Meta AI dominates technical problem-solving.

2. Subscription Models and Features

  • 2-1. ChatGPT Plus Subscription Details

  • ChatGPT Plus, provided by OpenAI, is an advanced subscription model available for $20 per month. The subscription includes access to the powerful GPT-4 model and the DALL-E 3 image generation tool. A unique and innovative feature of ChatGPT Plus is the GPT store, which allows users to build and share custom versions of ChatGPT optimized for various situations. However, unlike Gemini Advanced, ChatGPT Plus does not offer ancillary perks such as cloud storage.

  • 2-2. Gemini Advanced Subscription Features

  • Google’s Gemini Advanced subscription also costs $20 per month and provides access to Google's top AI model, Gemini Ultra 1.0. This package is inclusive of Google One’s offerings, which add significant value by providing 2 terabytes of cloud storage. Additionally, the subscription is expected to include Gemini integration for Gmail and Google Docs in the near future. Gemini Advanced stands out by packaging these additional benefits, which are not available in the standard free version. Google has also announced plans for a new model, Gemini Pro 1.5, which promises enhanced data processing capabilities, although it is not yet available to the public.

3. Performance Comparison: Google Gemini vs OpenAI ChatGPT

  • 3-1. Coding Proficiency

  • In a test comparing Google Gemini and OpenAI's ChatGPT for coding proficiency, both models were tasked with creating a Python script for a personal expense tracker. The evaluation criteria included functionality, ease of interaction, readability, and adherence to coding standards. While both chatbots successfully created functional scripts, Google Gemini demonstrated superior performance by adding extra functionality such as labels within a category and more granular reporting options. Therefore, Gemini was the winner in this category.

  • 3-2. Natural Language Understanding

  • To assess Natural Language Understanding (NLU), both chatbots were given a Cognitive Reflection Test question: 'A bat and a ball cost £1.10 in total. The bat costs £1.00 more than the ball. How much does the ball cost?' Both models correctly responded that the ball costs 5 cents, but OpenAI's ChatGPT was judged the winner due to its clearer explanation and detailed step-by-step solution.

  • 3-3. Creative Text Generation

  • For creative text generation, the chatbots were asked to write a short story set in a futuristic city where technology controls every aspect of life, but the protagonist discovers a hidden society living without modern tech. Both stories were original and adhered to the provided rubric. However, Google Gemini was rated higher overall for better narrative consistency and adherence to the theme, making it the winner in this category.

  • 3-4. Reasoning and Simplification

  • The chatbots were tasked with solving the classic 'two doors' problem: 'You are facing two doors. One door leads to safety, the other to danger. There are two guards, one always tells the truth, and the other always lies. You can ask one guard one question to determine the safe door. What question do you ask?' Both chatbots provided the correct question: 'Which door would the other guard say leads to danger?' However, OpenAI's ChatGPT was the winner due to its more detailed and clearer explanation.

  • 3-5. Ethical Decision-Making

  • For ethical decision-making, both chatbots were given a scenario involving an autonomous vehicle's moral decision: 'Should it hit a pedestrian or swerve and risk the lives of its passengers?' Neither chatbot provided a direct opinion but outlined various ethical frameworks and perspectives. Google Gemini was considered to have a more nuanced response with careful consideration of the various factors, and this was confirmed by a separate blind test using multiple models, solidifying Gemini's win.

  • 3-6. Translation

  • The task involved translating an English paragraph about U.S. Thanksgiving to French while emphasizing cultural nuances. Both chatbots performed well, but Google Gemini provided a more nuanced translation and explained its approach, thereby winning this category.

4. Performance Comparison: Meta AI vs ChatGPT vs Google Gemini

  • 4-1. Email Generation

  • Each AI chatbot, namely Meta AI, ChatGPT, and Google Gemini, was tasked with writing an email to request a project extension. All three created well-written, polite, and professional emails in a template style, allowing for personalization. This task was executed perfectly by all three, with no discernible differences in performance.

  • 4-2. Recipe Provision

  • When asked to provide a recipe for chili, Meta AI and Google Gemini both included sources for their recipes, with Gemini offering links to additional recipes. ChatGPT, on the other hand, failed to provide any sourcing, raising concerns about the accuracy and safety of its recipe. Therefore, Meta AI and Google Gemini are preferable for recipe provisions due to their traceable and verified sources.

  • 4-3. News Summarization

  • All three chatbots effectively provided a bulleted list of the latest news. Meta AI and ChatGPT linked directly to the news sources they cited, enhancing their reliability. In contrast, Google Gemini only mentioned various news sites without direct links, making it less useful for verification purposes. Thus, Meta AI and ChatGPT are more reliable for news summarization.

  • 4-4. Math Problem-Solving

  • In solving an algebra and a geometry problem, Meta AI emerged as the most reliable, providing accurate solutions to both problems. ChatGPT struggled and failed to provide a final result for the geometry problem, while Gemini offered theoretical answers without numeric values. Consequently, Meta AI is the most dependable for math problem-solving.

  • 4-5. Programming Tasks

  • For a programming prompt involving a variant of the tic-tac-toe game, Meta AI and ChatGPT successfully provided complete code in both HTML and JavaScript. Google Gemini, however, provided JavaScript code but substituted HTML with CSS, which is not interchangeable. Therefore, Meta AI and ChatGPT are the most reliable for programming tasks.

  • 4-6. Mock Interviews

  • In conducting a mock interview for a role as a computing staff writer, all three chatbots simulated comprehensive interviews with mock questions and answers. While each approached the task differently, all three provided useful starting points for understanding potential interview scenarios. Hence, Meta AI, ChatGPT, and Google Gemini all performed well in this area.

5. Key Findings and Practical Implications

  • 5-1. Strengths and Weaknesses

  • The analysis conducted across multiple tests demonstrated distinct strengths and weaknesses for each AI chatbot. ChatGPT, powered by GPT-4, impressed with its reasoning capabilities and the clarity of its problem-solving explanations, particularly in situations that involved ambiguity or logical puzzles. Gemini, offered by Google, excelled in creative text generation, coding capabilities, and natural language understanding. Its granular reporting options in coding tasks and its creative storytelling adherence to themes were notable strengths. Meta AI was found to be the most reliable in terms of overall performance, particularly in solving complex math problems and providing thorough programming code. However, all chatbots showed weaknesses as well. ChatGPT, despite its clear explanations, sometimes faltered in providing accurate sources for information, especially in recipe generation. Gemini, while excellent in certain technical tasks, lacked consistency in other areas such as handling sarcasm in conversational contexts. Meta AI, although strong in technical tasks, faced challenges in sourcing information reliably.

  • 5-2. Best Use Cases for Each Chatbot

  • Based on the comparative analysis, the best use cases for each AI chatbot are evident. ChatGPT is best suited for tasks that require clarity in reasoning and problem-solving, such as ethical decision-making and explaining complex concepts in a simple manner. It's also beneficial for generating summaries and engaging in deep analytical tasks. Google’s Gemini shines in scenarios requiring creativity and detailed coding assistance. It is ideal for generating creative content, writing stories, and tackling detailed programming challenges. Moreover, its image generation and superior natural language understanding make it a useful tool for creative professionals. Meta AI stands out as the most versatile tool, particularly excelling in mathematical problem-solving and reliable programming outputs. It is highly suitable for technical tasks, educational purposes, and ensuring accuracy in information retrieval. Its ability to handle everyday tasks with high consistency makes it an excellent all-rounder for general use.

6. Conclusion

  • The comparative analysis underscores that each AI chatbot—OpenAI's ChatGPT, Google's Gemini, and Meta's AI—has unique strengths and suits different tasks. ChatGPT, with its clear reasoning and problem-solving skills, is ideal for ethical decision-making and analytical tasks. Google Gemini is superior in creative text generation and coding proficiency, making it excellent for content creation and programming challenges. Meta AI stands out for its reliability in math and programming tasks, proving to be versatile and useful for technical and everyday functions. However, the report acknowledges the limitations and evolving nature of AI, suggesting continuous evaluations to stay abreast of advancements. Users are encouraged to choose a chatbot based on specific needs and leverage each tool's strengths for optimal results.

7. Glossary

  • 7-1. ChatGPT Plus [Subscription Service]

  • A subscription plan offered by OpenAI for ChatGPT, priced at $20 per month. It includes features such as creating custom versions via GPT store and offers advanced functionalities compared to the free version.

  • 7-2. Gemini Advanced [Subscription Service]

  • Google's subscription plan for its AI chatbot, Gemini, featuring additional cloud storage and other enhanced features. Priced similarly to ChatGPT Plus at $20 per month.

  • 7-3. Meta AI [AI Chatbot]

  • An AI chatbot developed by Meta that excels in tasks like providing complete answers to math problems and generating source-cited content, distinguishing itself through reliability and thoroughness.

  • 7-4. Coding Proficiency [Performance Metric]

  • A benchmark used to assess the ability of AI chatbots to write correct and efficient code. In the evaluations, Gemini showed superior coding proficiency.

  • 7-5. Creative Text Generation [Performance Metric]

  • A measure of how well AI chatbots can generate creative written content. Google's Gemini demonstrated strong performance in this area compared to other chatbots.

  • 7-6. Ethical Decision-Making [Performance Metric]

  • A task assessing the ability of AI chatbots to provide nuanced ethical reasoning. Gemini offered more insightful responses in this category in the comparative tests.

8. Source Documents