z-logo
Discover

Let's Analyze What Makes a Good H-Index Score

calendarMar 7, 2023 |clock8 Mins Read

Understanding H-Index

The H-index is a metric that measures an author’s productivity by the number of publications that have published their work and the impact of the work based on the number of citations their research receives. In general, authors with a higher h-index score will have produced more research and therefore published more content which, to their peers, creates their reputation of credibility.

This quantitative metric was brought about in 2005 by Argentinian-American professor of physics Jorge E. Hirsch to analyse publication data.

Finding an author’s H-index

There are multiple platforms on which you may find an author’s H-index score. To name a few, Google Scholar, Scopus, Web of science etc. However, in this blog, we’ll take you through the process of locating an author’s H-index on google scholar as shown below.

  1. Visit Google Scholar
  2. Enter the author’s name in the search bar
  3. If a profile exists for the author, it will appear at the top of the search results, click the author's name, and their profile page will open.
  4. View their h-index on the right side of the screen.

Calculating H-Index Score

The H-index measures the importance, significance, and impact of research contributions. To calculate an author’s H-index, you’d need to create a list of all publications in which the author has been published and rank them in descending order of the citations his/her work has received. Understanding the H-index of an author is an indication of their credibility, so that brings us to the question:

What is a good H-index score?

J. E. Hirsch (2005) observes that Noble Prize winners in physics have an average H-index score of 30, this highlights that Noble prize winners are selected with a scientific body of research and a history of contributional impact. This proved that successful scientists do need a good h-index score.

Hirsch stated that after 20 years of research; an H-index score of 20 was good, 40 was outstanding and 60 was truly exceptional. 

Does the H-index score evaluate an author in all important aspects?

Undoubtedly, it is appealing to have a singular value that measures an author's productivity and impact. Many committees have opted the H-index as their metric of choice as well. Bordons and Costas (2007) stated that the key advantage of the H-index metric is that it measures the scientific output of a researcher with objectivity. This plays a vital role in making decisions about promotions, fund allocation and awarding prizes.

However, there are suggestions that H-index does not take other important variables into account. According to Enago Academy (2022), a higher H-index score does not indicate better quality of research. The article further elaborates that the H-index score does not account for an author’s career stage, research and journal quality and contribution to the scientific community. The score also has potential unintended negative impacts; for example, a younger researcher may not challenge a researcher with a high h-index score and researchers aiming for a higher h-index may only pursue popular fields of science.

Furthermore, BiteSizeBio (2021) states that the H-index score does not take into account the number of authors on a research paper. If a paper has 1 author with about 100 citations, this researcher deserves more recognition than a paper that had 10 authors with similar citations. 

The fluctuation of the H-index score

The H-index score does not decrease unless the paper is redacted or deleted. Older papers may continue to gain new citations, and the h-index can potentially increase indefinitely, even after the researcher has stopped actively publishing. 

What is the difference between H-index and the journal impact factor?

The Journal Impact Factor metric is used to evaluate the importance of a journal within its respective field or discipline. In simpler terms, it measures the frequency of citations the average article within this journal receives. On the contrary, the H-index metric is used to measure the productivity and quality of an author’s publications. While they are both measures of research quality, they measure different aspects of research and can therefore not be compared.

To conclude, having a good H-index score is impressive. However, every author’s research contrasts with that of another. There are many more aspects to investigate when evaluating a researcher.

Discover millions of e-books, journal articles, proceedings and more on Zendy.

You might also like
Top 5 AI Ethical Issues that Can Impact Your Research Integrity
Jan 10, 20257 Mins ReadDiscover

Top 5 AI Ethical Issues that Can Impact Your Research Integrity

In a recent blog, we discussed responsible AI in research and why it matters. Now, we’ll discuss some AI ethical issues and what you should not be doing with AI in your research journey. This blog looks at common mistakes people make with AI in research, explains why they happen, and offers practical tips to avoid them. 1. Trusting AI Outputs Without Checking Them One big AI ethical issue is trusting everything AI tools generate without taking the time to verify it. AI models like ChatGPT can produce convincing answers, but they’re not always accurate. In research, this can lead to spreading incorrect information or drawing the wrong conclusions. Why It Happens: AI systems learn from existing data, which might include errors or biases. As a result, they can unintentionally repeat those issues. What You Can Do: Treat AI-generated content as a helpful draft, not the final word. Always double-check the information with reliable sources. 2. Using AI for Tasks That Require Human Judgment Relying on AI for decisions that need a human touch, like reviewing academic papers, is risky. These tasks often require context and empathy, which AI doesn’t have. Why It Happens: AI seems efficient, but it doesn’t understand the subtleties of human situations, leading to potential AI ethical issues in judgment and fairness. What You Can Do: Let AI assist with organizing or summarizing information, but make sure a person is involved in decisions that affect others. 3. Not Giving Credit to AI Tools Even when AI is used responsibly, failing to acknowledge its role can mislead readers about the originality of your work. Why It Happens: People might not think of AI as a source that needs to be cited, overlooking important AI ethical issues related to transparency and attribution. What You Can Do: Treat AI tools like any other resource. Check your institution’s or publisher’s guidelines for how to cite them properly. 4. Over-Reliance on AI for Creative Thinking AI can handle repetitive tasks, but depending on it too much can stifle human creativity. Research often involves brainstorming new ideas, which AI can’t do as well as people. Why It Happens: AI makes routine tasks more manageable, so letting it take over more complex ones is tempting. What You Can Do: Use AI to free up critical thinking and creative problem-solving time. Let it handle the busy work while you focus on the bigger picture to avoid these AI ethical issues. 5. Giving AI Access to Sensitive Data Allowing AI tools to access personal information without proper permission can pose serious security risks. Why It Happens: Some AI tools require access to data to function effectively, but their security measures might not be sufficient leading to potential AI ethical issues. What You Can Do: Limit the data AI tools can access. Use platforms with strong security features and comply with data protection regulations. Final Thoughts AI can be a valuable tool for researchers, but it’s not without its challenges. Many of these challenges stem from AI ethical issues that arise when AI is misused or misunderstood. By understanding these common mistakes and taking steps to address them, you can use AI responsibly and effectively. The key is to see AI as an assistant that complements human effort, not a replacement. .wp-block-image img { max-width: 80% !important; margin-left: auto !important; margin-right: auto !important; }

Speed Up Your Research With “Insights”
Dec 18, 20243 Mins ReadDiscover

Speed Up Your Research With “Insights”

'Insights', a brand-new feature designed to make your research experience faster, simpler, and more accessible. Insights gives you short, clear summaries of research papers, pulling out the most important information so you can understand the main points in just a few lines. Instead of reading through pages of dense content, you’ll get a quick overview that helps you decide if the paper is worth exploring further. Here’s how Insights can help: Save time by getting to the heart of a paper faster. Understand complex topics without feeling stuck. Focus on what matters and decide quickly what’s relevant to you. Why We Created Insights? We’ve heard from many of you that keeping up with research can feel like a never-ending task. There’s so much to read, and it’s hard to know where to start. That’s where 'Insights' comes in, to help you make the most of your time exploring the right research paper you are looking for. How Does It Work? Insights uses our AI to scan through a paper and extract key points. It focuses on sections like the introduction, methodology, results, and conclusion, so you can get a clear sense of what the paper is about. You don’t have to worry about missing anything important; it’s all laid out in a simple, easy-to-digest format. Head over to Zendy, search for what you are looking for, and see how Insights can give you a clearer overview in seconds, Check out Insights now! .wp-block-image img { max-width: 65% !important; margin-left: auto !important; margin-right: auto !important; }

Responsible AI In Research And Why It Matters
Dec 18, 20249 Mins ReadDiscover

Responsible AI In Research And Why It Matters

Artificial Intelligence (AI) is changing how we live, work, and learn. However, as AI continues to evolve, it is important to ensure it is developed and used responsibly. In this blog, we’ll explore what responsible AI means, why it is essential, and how tools like ZAIA, Zendy's AI assistant for researchers, implement these principles in the academic sector. What Is Responsible AI? Responsible AI, also known as ethical AI refers to building and using AI tools guided by key principles: Fairness Reliability Safety Privacy and Security Inclusiveness Transparency Accountability AI vs Responsible AI: Why Does Responsible AI Matter? Keep in mind that AI is not a human being. This means it lacks the ability to comprehend ethical standards or a sense of responsibility in the same way humans do. Therefore, ensuring these concepts are embedded in the development team before creating the tool is more important than building the tool itself. In 2016, Microsoft launched a Twitter chatbot called "Tay", a chatbot designed to entertain 18- to 24-year-olds in the US to explore the conversational capabilities of AI. Within just 16 hours, the tool's responses turned toxic, racist, and offensive due to being fed harmful and inappropriate content by some Twitter users. This led to the immediate shutdown of the project, followed by an official apology from the development team. In such cases, "Tay" lacked ethical guidelines to help it differentiate harmful content from appropriate content. For this reason, it is crucial to train AI tools on clear principles and ethical frameworks that enable them to produce more responsible outputs.The development process should also include designing robust monitoring systems to continuously review and update the databases' training, ensuring they remain free of harmful content. Overall, the more responsible the custodian is, the better the child’s behaviour will be. The Challenges And The Benefits of Responsible AI Responsible AI is not a "nice-to-have" feature, it’s a foundational set of principles that every AI-based tool must implement. Here's why: Fairness: By addressing biases, responsible AI ensures every output is relevant and fair for all society’s values. Trust: Transparency in how AI works builds trust among users. Accountability: Developers and organisations adhere to high standards, continuously improving AI tools and holding themselves accountable for their outcomes. This ensures that competition centers on what benefits communities rather than simply what generates more revenue. Implementing responsible AI comes with its share of challenges: Biased Data: AI systems often learn from historical data, which may carry biases. This can lead to skewed outcomes, like underrepresenting certain research areas or groups. Awareness Gaps: Not all developers and users understand the ethical implications of AI, making education and training critical. Time Constraints: AI tools are sometimes developed rapidly, bypassing essential ethical reviews, which increases the risk of errors. Responsible AI and ZAIA ZAIA, Zendy’s AI-powered assistant for researchers, is built with a responsible AI framework in mind. Our AI incorporates the six principles of responsible AI, fairness, reliability and safety, privacy and security, inclusiveness, transparency, and accountability, to meet the needs of students, researchers, and professionals in academia. Here’s how ZAIA addresses these challenges: Fairness: ZAIA ensures balanced and unbiased recommendations, analysing academic resources from diverse disciplines and publishers. Reliability and Safety: ZAIA’s trained model is rigorously tested to provide accurate and dependable insights, minimising errors in output. Transparency: ZAIA’s functionality is clear and user-friendly, helping researchers understand and trust its outcomes. Accountability: Regular updates improve ZAIA’s features, addressing user feedback and adapting to evolving academic needs. Conclusion Responsible AI is the foundation for building ethical and fair systems that benefit everyone. ZAIA is Zendy’s commitment to this principle, encouraging users to explore research responsibly and effectively. Whether you’re a student, researcher, or professional, ZAIA provides a reliable and ethical tool to enhance your academic journey. Discover ZAIA today. Together, let’s build a future where AI serves as a trusted partner in education and beyond. .wp-block-image img { max-width: 80% !important; margin-left: auto !important; margin-right: auto !important; }