Overview of Bibliometrics - IAP Course version 1.1Micah Altman
Whose articles cite a body of work? Is this a high-impact journal? How might others assess my scholarly impact? Citation analysis is one of the primary methods used to answer these questions.
Data Publishing Workflows with DataverseMicah Altman
By: Mercè Crosas, Director of Data Science at the Institute for Quantitative Social Science (IQSS) at Harvard University
The Dataverse software provides multiple workflows for data publishing to support a wide range of data policies and practices established by journals, as well as data sharing needs from various research communities. This talk will describe these workflows from the user experience and from the system's technical implementation.
This talk was presented as part of the Information Science Brown Bag talks, hosted by the Program on Information Science. (See http://drmaltman.wordpress.com)
Using Bibliometrics Tools to Increase the visibility of your publicationsCiarán Quinn
Strategies to increase the visibility of your research including using keywords, Bibliometric resources, measuring your H Index,Journal Impact, Article level metrics, Altmetrics, and Academic Social Networks
Google Scholar Citations allows researchers to track citations to their publications. It provides metrics like citations over time and who is citing an author's work. Setting up a profile is quick and easy - researchers can login with their Google account, add their publications, and make their profile public. Profile and citation records are automatically updated as Google Scholar indexes new information. The tool helps authors explore citations, follow new citations, and set alerts to new references to their work.
Publish or Perish is a free software program that uses data from Google Scholar to provide citation metrics and statistics about researchers' work, including the total number of papers and citations, the h-index and g-index. It allows you to search for authors to analyze their impact. The h-index quantifies both productivity and citation impact by calculating the number of papers with at least that number of citations.
This document provides a summary of a presentation on bibliometrics. It discusses the basics of bibliometrics including who cites whom and patterns of scholarly research. It then covers major citation databases like Web of Science, Scopus, and Google Scholar and compares their coverage. Free online sources of bibliometric data are also presented including Eigenfactor, SJR, and Publish or Perish. Emerging altmetrics and new bibliometric indicators are discussed. The future of bibliometrics and how to follow new developments are addressed at the end.
Scientometrics is the quantitative study of science, communication in science, and science policy. It involves the analysis, evaluation, and visualization of science using quantitative methods. Some key points:
- The term was coined in 1969 in Russia and gained recognition after the founding of the journal Scientometrics in 1978.
- It applies bibliometric methods to measure the impact and quality of science by analyzing patterns of citation and publication.
- Important indicators include the journal impact factor, h-index, and citation analysis which are used to evaluate scientific institutions, departments, and individual researchers.
- Tools like CiteSpace and Authormap are used to map and visualize citation networks and trends within literature.
Overview of Bibliometrics - IAP Course version 1.1Micah Altman
Whose articles cite a body of work? Is this a high-impact journal? How might others assess my scholarly impact? Citation analysis is one of the primary methods used to answer these questions.
Data Publishing Workflows with DataverseMicah Altman
By: Mercè Crosas, Director of Data Science at the Institute for Quantitative Social Science (IQSS) at Harvard University
The Dataverse software provides multiple workflows for data publishing to support a wide range of data policies and practices established by journals, as well as data sharing needs from various research communities. This talk will describe these workflows from the user experience and from the system's technical implementation.
This talk was presented as part of the Information Science Brown Bag talks, hosted by the Program on Information Science. (See http://drmaltman.wordpress.com)
Using Bibliometrics Tools to Increase the visibility of your publicationsCiarán Quinn
Strategies to increase the visibility of your research including using keywords, Bibliometric resources, measuring your H Index,Journal Impact, Article level metrics, Altmetrics, and Academic Social Networks
Google Scholar Citations allows researchers to track citations to their publications. It provides metrics like citations over time and who is citing an author's work. Setting up a profile is quick and easy - researchers can login with their Google account, add their publications, and make their profile public. Profile and citation records are automatically updated as Google Scholar indexes new information. The tool helps authors explore citations, follow new citations, and set alerts to new references to their work.
Publish or Perish is a free software program that uses data from Google Scholar to provide citation metrics and statistics about researchers' work, including the total number of papers and citations, the h-index and g-index. It allows you to search for authors to analyze their impact. The h-index quantifies both productivity and citation impact by calculating the number of papers with at least that number of citations.
This document provides a summary of a presentation on bibliometrics. It discusses the basics of bibliometrics including who cites whom and patterns of scholarly research. It then covers major citation databases like Web of Science, Scopus, and Google Scholar and compares their coverage. Free online sources of bibliometric data are also presented including Eigenfactor, SJR, and Publish or Perish. Emerging altmetrics and new bibliometric indicators are discussed. The future of bibliometrics and how to follow new developments are addressed at the end.
Scientometrics is the quantitative study of science, communication in science, and science policy. It involves the analysis, evaluation, and visualization of science using quantitative methods. Some key points:
- The term was coined in 1969 in Russia and gained recognition after the founding of the journal Scientometrics in 1978.
- It applies bibliometric methods to measure the impact and quality of science by analyzing patterns of citation and publication.
- Important indicators include the journal impact factor, h-index, and citation analysis which are used to evaluate scientific institutions, departments, and individual researchers.
- Tools like CiteSpace and Authormap are used to map and visualize citation networks and trends within literature.
This document discusses building citation indexes from different databases including Scopus, Web of Science (WoS), and CSA (Cambridge Scientific Abstracts). It describes algorithms used to match cited references to full bibliographic records within and across these databases. For Scopus and WoS, high matching rates of 90% were achieved, while for CSA (social sciences) the matching rate was lower at 30% due to references being to non-journal materials like books. The document outlines challenges in citation indexing for social sciences and a vision for a distributed, semantic citation index based on full-text collections and reference linking.
Towards a Semantic Citation Index for the German Social SciencesGESIS
The document discusses building a semantic citation index for German social sciences research. It notes the inadequacies of existing citation indexes in representing this research. It outlines initial steps taken, including matching references in CSA databases to build connections between records. Future plans include expanding text collections, developing natural language processing methods to extract additional semantic information from citations, and creating an open and distributed citation index. The long term goal is a globally networked citation index to support bibliometric analysis of social sciences literature.
Scientometric Mapping of Library and Information Science in Web of Science 8638812142
This is a presentation on Scientometric Study done in Library and Information Science Research as per the data downloaded from Web of Science. This is a presentation of MPhil dissertation submitted to Department of Library and Information Science, Mizoram University under Prof SN Singh.
1) The document describes the SOPHIA project, which aims to build altmetric networks of researchers and institutions to understand how research impacts spread in society.
2) SOPHIA collects data from Scopus and social media sources to build a heterogeneous graph network, and analyzes the network using graph metrics to measure the influence and authority of researchers and institutions.
3) The project has developed visualization and search tools to explore the altmetric networks, annotated documents, and metrics within a software prototype called SOPHIA.
Profiling systems have achieved notable adoption by research institutions.1 Multi-site search of research profiling systems has substantially evolved since the first deployment of systems such as DIRECT2Experts.2 CTSAsearch is a federated search engine using VIVO-compliant Linked Open Data (LOD) published by members of the NIH-funded Clinical and Translational Science (CTSA) consortium and other interested parties. Sixty-four institutions are currently included, spanning six distinct platforms and three continents (North America, Europe and Australia). In aggregate, CTSAsearch has data on 150-300 thousand unique researchers and their 10 million publications. The public interface is available at http://research.icts.uiowa.edu/polyglot.
Lecture presented by Marian Ramos Eclevia at PAARL's Summer Conference on the theme "Library Analytics: Data-driven Library Management", held at Pearl Hotel, Manila on 20-22 April 2016
This document provides an overview of resources for researching the impact of library usage on GPA using sociological databases and methods. It discusses using SocIndex, Sociological Abstracts, Psycinfo and ERIC to explore related topics, variables, and disciplines involved. It also addresses using quantitative, qualitative or mixed research methods. Finally, it provides tips for searching databases and formatting references in ASA style.
This 90 minutes workshop is the first part of the library training series, designed to enhance the library knowledge of the front-line support staff in basic search skills.
Research Impact in Specialized Settings: 3 Case StudiesElaine Lasda
Presentation of 3 case studies where research impact metrics are used to further the mission of institutions and organizations out of the traditional academic millieu.
Resources in uct libraries org_psych_m_philpeoplemanment_goodman_2015Susanne Noll
This document provides an introduction to resources available in the University of Cape Town (UCT) Libraries to assist with literature reviews. It discusses print and electronic resources such as journals, databases, and RefWorks. The presentation teaches how to search the library catalog to find books, access electronic journals, use bibliographic databases like EbscoHost and Google Scholar to find articles, and utilize the citation management tool RefWorks. Contact information for research help from librarians is also provided.
This presentation is about shortlisting and choosing journals for publishing. It also discusses quality issues, including predatory and hijacked journals. Most appropriate for Social Science students.
Searching beyond datasets in the Social SciencesGESIS
1) State of the art search capabilities at GESIS include curated datasets with metadata and additional materials as well as harvested datasets and linking infrastructure between items.
2) Integrated search allows linking of papers and datasets in both directions to help users.
3) However, social scientists often want to reuse specific parts of datasets, like questions on a particular topic, requiring more fine-grained retrievability and linkages within datasets.
This document provides guidance for students on conducting research and citing sources for a project through the UConn and HK Library websites. It instructs students to visit these library websites to access databases, books, and other resources. The document outlines MLA citation format and provides examples of in-text and reference page citations. It also provides guidance on finding and citing images, as well as citing sources found through databases.
The New Metrics: conference presentationElaine Lasda
This document discusses innovative uses of research impact indicators and metrics. It provides examples of how research institutions like the University of Michigan Publishing, EPA Research Triangle Park, and the National Center for Atmospheric Research have used metrics to demonstrate the broader impact and value of research to various stakeholders. It also outlines some of the shared challenges these institutions face in gathering and contextualizing impact data, as well as opportunities for librarians to play a leadership role in these efforts through skills in project management, data analysis, and relationship building. Overall, the document argues that understanding and communicating research impact can help validate funding and build partnerships across organizations.
This document discusses various metrics used to evaluate academic journals and research, including impact factor, H-index, indexing, citations, and open access publishing. It describes what academic journals are and their purposes of peer review and sharing new research. It also discusses the rise of predatory journals and how to identify them, as well as advantages and disadvantages of open access publishing. Key metrics for evaluating journals include impact factor, eigenfactor, SJR, and indexing in services like PubMed and Scopus. Metrics for evaluating individual researchers include H-index and citations. The document emphasizes that quality research should be the priority over metrics and impact.
Fixing the infrastructure for open scienceBjörn Brembs
Björn Brembs argues that the current system of scholarly publishing and metrics is dysfunctional. It limits access to literature, data, and software and does not allow for flexible data visualization, analysis, or discovery. Journal impact factors in particular are mathematically unsound and an unreliable measure of quality. Better alternatives are needed that provide open access to all scholarly outputs and allow literature, data, and software to be easily searched, accessed, analyzed, and built upon.
V.3 poster current citations and a future with linked dataIliadis Dimitrios
1) Converting citation data to linked data has several advantages such as allowing other applications to use the citation data, describing the reasons publications were cited, and connecting citation information like authors and papers.
2) Linked data assigns unique identifiers (URIs) to citations and related information and describes relationships between cited and citing publications using RDF triples. This allows connecting citation data to other linked open data.
3) Projects that convert citation data to linked data use URIs, RDF triples, and ontologies like CiTO to describe citation intent. This enables advanced searches, citation network visualizations, and linking to other semantic data.
This document provides an overview of information resources for conducting research. It discusses developing a focused research question, identifying relevant terms and databases to search, evaluating search results, and tools for managing citations. Key databases recommended for answering public health and ethnic minority questions include Medline, Embase, and CINAHL. The document emphasizes refining searches and stresses that librarians can help at all stages of the research process.
This document discusses the role of libraries in research evaluation. It provides an overview of research evaluation in the UK context and how bibliometric measures like the number of publications, citations, h-index, and journal impact factors are used. It explains data sources like Web of Science, Scopus and Google Scholar and how to interpret citation data. The document also discusses how libraries can advise researchers on using bibliometric data to tell the story of their research impact and provides examples of high cited articles and institutional rankings. It proposes topics for librarians to discuss with researchers to help them understand research evaluation and metrics.
This document provides an overview of a workshop on publishing and getting research read. It discusses various topics related to the publication process, including why researchers publish, finding the best journal to submit to, open access publishing, how to interpret impact factors, writing manuscripts, publication ethics, and peer review. The document also includes an introduction to the workshop presenter, Max Haring, who is an executive editor at Springer and has a PhD in plant epigenetics.
This document discusses building citation indexes from different databases including Scopus, Web of Science (WoS), and CSA (Cambridge Scientific Abstracts). It describes algorithms used to match cited references to full bibliographic records within and across these databases. For Scopus and WoS, high matching rates of 90% were achieved, while for CSA (social sciences) the matching rate was lower at 30% due to references being to non-journal materials like books. The document outlines challenges in citation indexing for social sciences and a vision for a distributed, semantic citation index based on full-text collections and reference linking.
Towards a Semantic Citation Index for the German Social SciencesGESIS
The document discusses building a semantic citation index for German social sciences research. It notes the inadequacies of existing citation indexes in representing this research. It outlines initial steps taken, including matching references in CSA databases to build connections between records. Future plans include expanding text collections, developing natural language processing methods to extract additional semantic information from citations, and creating an open and distributed citation index. The long term goal is a globally networked citation index to support bibliometric analysis of social sciences literature.
Scientometric Mapping of Library and Information Science in Web of Science 8638812142
This is a presentation on Scientometric Study done in Library and Information Science Research as per the data downloaded from Web of Science. This is a presentation of MPhil dissertation submitted to Department of Library and Information Science, Mizoram University under Prof SN Singh.
1) The document describes the SOPHIA project, which aims to build altmetric networks of researchers and institutions to understand how research impacts spread in society.
2) SOPHIA collects data from Scopus and social media sources to build a heterogeneous graph network, and analyzes the network using graph metrics to measure the influence and authority of researchers and institutions.
3) The project has developed visualization and search tools to explore the altmetric networks, annotated documents, and metrics within a software prototype called SOPHIA.
Profiling systems have achieved notable adoption by research institutions.1 Multi-site search of research profiling systems has substantially evolved since the first deployment of systems such as DIRECT2Experts.2 CTSAsearch is a federated search engine using VIVO-compliant Linked Open Data (LOD) published by members of the NIH-funded Clinical and Translational Science (CTSA) consortium and other interested parties. Sixty-four institutions are currently included, spanning six distinct platforms and three continents (North America, Europe and Australia). In aggregate, CTSAsearch has data on 150-300 thousand unique researchers and their 10 million publications. The public interface is available at http://research.icts.uiowa.edu/polyglot.
Lecture presented by Marian Ramos Eclevia at PAARL's Summer Conference on the theme "Library Analytics: Data-driven Library Management", held at Pearl Hotel, Manila on 20-22 April 2016
This document provides an overview of resources for researching the impact of library usage on GPA using sociological databases and methods. It discusses using SocIndex, Sociological Abstracts, Psycinfo and ERIC to explore related topics, variables, and disciplines involved. It also addresses using quantitative, qualitative or mixed research methods. Finally, it provides tips for searching databases and formatting references in ASA style.
This 90 minutes workshop is the first part of the library training series, designed to enhance the library knowledge of the front-line support staff in basic search skills.
Research Impact in Specialized Settings: 3 Case StudiesElaine Lasda
Presentation of 3 case studies where research impact metrics are used to further the mission of institutions and organizations out of the traditional academic millieu.
Resources in uct libraries org_psych_m_philpeoplemanment_goodman_2015Susanne Noll
This document provides an introduction to resources available in the University of Cape Town (UCT) Libraries to assist with literature reviews. It discusses print and electronic resources such as journals, databases, and RefWorks. The presentation teaches how to search the library catalog to find books, access electronic journals, use bibliographic databases like EbscoHost and Google Scholar to find articles, and utilize the citation management tool RefWorks. Contact information for research help from librarians is also provided.
This presentation is about shortlisting and choosing journals for publishing. It also discusses quality issues, including predatory and hijacked journals. Most appropriate for Social Science students.
Searching beyond datasets in the Social SciencesGESIS
1) State of the art search capabilities at GESIS include curated datasets with metadata and additional materials as well as harvested datasets and linking infrastructure between items.
2) Integrated search allows linking of papers and datasets in both directions to help users.
3) However, social scientists often want to reuse specific parts of datasets, like questions on a particular topic, requiring more fine-grained retrievability and linkages within datasets.
This document provides guidance for students on conducting research and citing sources for a project through the UConn and HK Library websites. It instructs students to visit these library websites to access databases, books, and other resources. The document outlines MLA citation format and provides examples of in-text and reference page citations. It also provides guidance on finding and citing images, as well as citing sources found through databases.
The New Metrics: conference presentationElaine Lasda
This document discusses innovative uses of research impact indicators and metrics. It provides examples of how research institutions like the University of Michigan Publishing, EPA Research Triangle Park, and the National Center for Atmospheric Research have used metrics to demonstrate the broader impact and value of research to various stakeholders. It also outlines some of the shared challenges these institutions face in gathering and contextualizing impact data, as well as opportunities for librarians to play a leadership role in these efforts through skills in project management, data analysis, and relationship building. Overall, the document argues that understanding and communicating research impact can help validate funding and build partnerships across organizations.
This document discusses various metrics used to evaluate academic journals and research, including impact factor, H-index, indexing, citations, and open access publishing. It describes what academic journals are and their purposes of peer review and sharing new research. It also discusses the rise of predatory journals and how to identify them, as well as advantages and disadvantages of open access publishing. Key metrics for evaluating journals include impact factor, eigenfactor, SJR, and indexing in services like PubMed and Scopus. Metrics for evaluating individual researchers include H-index and citations. The document emphasizes that quality research should be the priority over metrics and impact.
Fixing the infrastructure for open scienceBjörn Brembs
Björn Brembs argues that the current system of scholarly publishing and metrics is dysfunctional. It limits access to literature, data, and software and does not allow for flexible data visualization, analysis, or discovery. Journal impact factors in particular are mathematically unsound and an unreliable measure of quality. Better alternatives are needed that provide open access to all scholarly outputs and allow literature, data, and software to be easily searched, accessed, analyzed, and built upon.
V.3 poster current citations and a future with linked dataIliadis Dimitrios
1) Converting citation data to linked data has several advantages such as allowing other applications to use the citation data, describing the reasons publications were cited, and connecting citation information like authors and papers.
2) Linked data assigns unique identifiers (URIs) to citations and related information and describes relationships between cited and citing publications using RDF triples. This allows connecting citation data to other linked open data.
3) Projects that convert citation data to linked data use URIs, RDF triples, and ontologies like CiTO to describe citation intent. This enables advanced searches, citation network visualizations, and linking to other semantic data.
This document provides an overview of information resources for conducting research. It discusses developing a focused research question, identifying relevant terms and databases to search, evaluating search results, and tools for managing citations. Key databases recommended for answering public health and ethnic minority questions include Medline, Embase, and CINAHL. The document emphasizes refining searches and stresses that librarians can help at all stages of the research process.
This document discusses the role of libraries in research evaluation. It provides an overview of research evaluation in the UK context and how bibliometric measures like the number of publications, citations, h-index, and journal impact factors are used. It explains data sources like Web of Science, Scopus and Google Scholar and how to interpret citation data. The document also discusses how libraries can advise researchers on using bibliometric data to tell the story of their research impact and provides examples of high cited articles and institutional rankings. It proposes topics for librarians to discuss with researchers to help them understand research evaluation and metrics.
This document provides an overview of a workshop on publishing and getting research read. It discusses various topics related to the publication process, including why researchers publish, finding the best journal to submit to, open access publishing, how to interpret impact factors, writing manuscripts, publication ethics, and peer review. The document also includes an introduction to the workshop presenter, Max Haring, who is an executive editor at Springer and has a PhD in plant epigenetics.
1. The document discusses preparing researchers for the next Research Excellence Framework (REF) assessment in the UK. It covers open access policies, bibliometrics, altmetrics, and ORCID identifiers.
2. Open access requirements for REF submissions are that journal articles and conference papers be made publicly available within 3 months of acceptance in an institutional repository.
3. Bibliometrics like citation counts and journal impact factors may play a larger role in REF assessments in the future, though peer review will still be primary. Concerns about gaming the system and disciplinary biases remain.
Research Data Access and Preservation Summit, 2016
Atlanta, GA
May 4-7, 2016
Presenters:
Abigail Goben, University of Illinois Chicago
Tina Griffin, University of Illinois Chicago
Sara Scheib, University of Iowa
Scott Martin, University of Michigan
Panel Leads:
Megan Sapp Nelson, Purdue University
Marina Zhang, University of Iowa
Academics must provide evidence to demonstrate the impact and outcomes of their scholarly work. This webinar, presented by librarians, will help faculty explore various forms of documentary evidence to support their case for excellence. Sponsored by the IUPUI Office of Academic Affairs.
Note: The webinar included demonstrations of Web of Science & Scopus, which the slides do not reflect.
1) The document discusses information use and user studies, including methods for studying print and electronic resource usage, information needs, information seeking behavior, and competencies needed by LIS professionals.
2) It provides details on various methods for conducting use studies, such as analyzing issue records, usage statistics, surveys and citations. It also outlines how to increase e-resource usage through marketing and outreach.
3) The document concludes by emphasizing the important role of LIS professionals in representing information to users, just as actors represent characters, through competencies like communication, computing, and information literacy skills.
Early Career Tactics to Increase Scholarly ImpactElaine Lasda
Workshp for Ph.D. candidates, postdocs and faculy on how bilbiometrics, altmetrics, open access, ORCID, and other resources enable greater visibility of research output.
The document introduces ARIA (Altmetrics for Research Impact Actuation), a prototype developed by NTU to measure research impact using altmetrics and bibliometrics. ARIA collects publication data and metrics from various sources for over 4,000 NTU researchers. It provides visualizations of metrics at researcher, school, and university levels over time. The document describes ARIA's data sources and extraction methodology, as well as screenshots of the researcher and admin dashboards, artifact pages, and cross-metric explorer feature. It also announces an upcoming altmetrics workshop at NTU in January 2018.
Academic Library Impact: Improving Practice and Essential Areas to ResearchLynn Connaway
The document discusses priority areas for researching the value and impact of academic libraries. It identifies the key areas as communication, mission alignment, learning analytics, student success, teaching and learning, and collaboration. For each area, it provides exemplar effective practices from literature and interviews with librarians and administrators. It then outlines potential research questions within each area and discusses research design considerations. The document concludes with an overview of a visualization tool being developed to showcase findings.
Academic Library Impact: Improving Practice and Essential Areas to ResearchOCLC
Connaway, Lynn Silipigni, William Harvey, Vanessa Kitzie, and Stephanie Mikitish. 2017. “Academic Library Impact: Improving Practice and Essential Areas to Research.” Presented at the Update on Value of Academic Libraries Initiative (ACRL) at the ALA Annual Conference, Chicago, Illinois, June 25.
Presented at the OCLC Research Library Partnership meeting by Senior Program Officer, Karen Smith-Yoshimura and hosted by the University of Sydney in Sydney, NSW Australia, 17 February 2017. This meeting provided an opportunity for Research Library Partners to touch base with each other on issues of common concern and explore possible areas of future engagement with the OCLC Research Library Partnership and OCLC Research.
This document provides an overview of bibliometrics and discusses various bibliometric indicators and tools. It describes what bibliometrics is, why it is used, and different bibliometric indicators like the impact factor, h-index, SNIP, SJR, and altmetrics. It discusses bibliometric data sources like Web of Science, Scopus, Google Scholar, and provides pros and cons of each. The document concludes that no single metric can provide a complete picture and that metrics should be used to improve research assessment rather than rely on a single number or tool.
Citation analysis is a method to measure the impact of research using bibliometric tools. It counts the number of times a researcher's published work has been cited by other researchers, with more citations indicating greater relevance. The h-index also measures impact by counting the number of papers cited at least h times. Citation analysis can be used for grant applications, promotion, benchmarking and identifying collaborators. However, tools vary by discipline and no single tool covers all publications, so it is important to be aware of limitations and use multiple tools. The document provides guidance on specific tools for different disciplines.
This document discusses bibliometrics and their use at Cardiff University. It begins with an introduction to bibliometric measures like citations, impact factors, and altmetrics. It then discusses how bibliometric data is presented in Cardiff's institutional repository and how it was used to provide context for research evaluations in the UK's REF2014 assessment exercise. The document concludes by outlining Cardiff's trial of the SciVal analytics tool and plans for a new research information system to better integrate bibliometric and altmetric data.
This document discusses various metrics for measuring the impact and importance of academic journals, articles, and authors. It describes journal impact factors, h-indexes, and other bibliometric tools like Web of Knowledge, Google Scholar, and Scopus that provide citation data. It notes that no single tool provides comprehensive coverage and that metrics can be influenced by many factors. The document cautions that impact metrics should not replace peer review and various limitations must be considered. It also introduces altmetrics that measure social media mentions as a new way to assess research impact.
Supporting Bibliometrics by Jenny Delasalle, Academic Support Manager (Research), University of Warwick. Presentation at the Research Evaluation: Is It Our Business? The Role of Librarians in the Brave New World of Research Evaluation 29 June 2011, University of Birmingham, Edgbaston Campus.
This document discusses resources for research and evaluating information sources. It provides guidance on searching library databases like Summon, which allows access to academic books and journals, versus Google. Key criteria are outlined for evaluating information sources, such as authority, relevance, objectivity and currency. Tips are provided for accessing materials not available in the library, including requesting books and using other libraries.
Selecting efficient and reliable preservation strategiesMicah Altman
This article addresses the problem of formulating efficient and reliable operational preservation policies that ensure bit-level information integrity over long periods, and in the presence of a diverse range of real-world technical, legal, organizational, and economic threats. We develop a systematic, quantitative prediction framework that combines formal modeling, discrete-event-based simulation, hierarchical modeling, and then use empirically calibrated sensitivity analysis to identify effective strategies.
This discussion, covened by the Dubai Future Foundation, focusses on identifying the significance of the concept of well-being for social-science and policy; and the opportunities to measure it at scale.
Matching Uses and Protections for Government Data Releases: Presentation at t...Micah Altman
In the work included below, and presented at the Simons Institute, we describe work-in progress that aims to align emerging methods of data protections with research uses.
Privacy Gaps in Mediated Library Services: Presentation at NERCOMP2019Micah Altman
Libraries enable patrons to access a wide range of information, but much of the access to this information is now directly managedy publishers. This has lead to a significant gap across library values, patrons perception of privacy, and effective privacy protection for access to digital resources.
In the work included below, and presented at NERCOMP 2019, we review privacy principles based on ALA, IFLA, and NISO policies. We then organizing and comparing high level privacy protections required by ALA checklist, NISO, and GDPR. This framework of principles and controls is then used to score the privacy policies and practices of major vendors of research library content. We evaluate each element of the vendors privacy policy, and use instrumented browsers to identify the types of tracking mechanisms used by different vendors. We use this set of privacy scores to support analyses of change over time, and of potential gaps between patron expectations and privacy policies and practices.
Presentation by Philip Cohen on collaborative work with Micah Altman as part of the MIT CREOS research talk series. Presented in fall 2018, in Cambridge, MA.
Contemporary journal peer review is beset by a range of problems. These include (a) long delay times to publication, during which time research is inaccessible; (b) weak incentives to conduct reviews, resulting in high refusal rates as the pace of journal publication increases; (c) quality control problems that produce both errors of commission (accepting erroneous work) and omission (passing over important work, especially null findings); (d) unknown levels of bias, affecting both who is asked to perform peer review and how reviewers treat authors, and; (e) opacity in the process that impedes error correction and more systematic learning, and enables conflicts of interest to pass undetected. Proposed alternative practices attempt to address these concerns -- especially open peer review, and post-publication peer review. However, systemic solutions will require revisiting the functions of peer review in its institutional context.
The document discusses peer review reform and proposes a new system called IOTA (I Owe the Academy Review). IOTA would allow academics to donate review tokens that represent a pledge to conduct peer reviews. These tokens could then be granted to various peer review initiatives in exchange for conducting experiments and sharing results openly. The goal is to better match reviewers with projects promoting high-quality scholarship, while generating more empirical evidence on effective peer review methods. Several example scenarios for how IOTA could work with different types of journals or research pools are provided.
Redistricting in the US -- An OverviewMicah Altman
This presentation was prepared for the International Seminar on Electoral Districting, National Electoral Institute El Colegio de México. http://www.ine.mx/seminario-internacional-distritacion-electoral/
This document summarizes a presentation about the future of electoral redistricting. It discusses how open data, transparent criteria, public participation, and open-source software can improve redistricting. Key points include:
- Open data is needed for evaluating proposals, modifying plans, and auditing the process. Keeping data open requires coordination.
- Criteria should be based on social science, validated, and allow for reproducibility and robustness against small changes.
- Public participation exists on a continuum from information seeking to proposing alternatives. It can increase engagement, legitimacy and accountability.
- Independent institutions are important to support criteria, funding, transparency and public participation in the process.
- Open-source software can enable
A History of the Internet :Scott Bradner’s Program on Information Science Talk Micah Altman
Scott Bradner is a Berkman Center affiliate who worked for 50 at Harvard in the areas of computer programming, system management, networking, IT security, and identity management. Scott Bradner was involved in the design, operation and use of data networks at Harvard University since the early days of the ARPANET and served in many leadership roles in the IETF. He presented the talk recorded below, entitled, A History of the Internet -- as part of Program on Information Science Brown Bag Series:
Bradner abstracted his talk as follows:
In a way the Russians caused the Internet. This talk will describe how that happened (hint it was not actually the Bomb) and follow the path that has led to the current Internet of (unpatchable) Things (the IoT) and the Surveillance Economy.
SAFETY NETS: RESCUE AND REVIVAL FOR ENDANGERED BORN-DIGITAL RECORDS- Program ...Micah Altman
The web is now firmly established as the primary communication and publication platform for sharing and accessing social and cultural materials. This networked world has created both opportunities and pitfalls for libraries and archives in their mission to preserve and provide ongoing access to knowledge. How can the affordances of the web be leveraged to drastically extend the plurality of representation in the archive? What challenges are imposed by the intrinsic ephemerality and mutability of online information? What methodological reorientations are demanded by the scale and dynamism of machine-generated cultural artifacts? This talk will explore the interplay of the web, contemporary historical records, and the programs, technologies, and approaches by which libraries and archives are working to extend their mission to preserve and provide access to the evidence of human activity in a world distinguished by the ubiquity of born-digital materials.
Information Science Brown Bag talks, hosted by the Program on Information Science, consists of regular discussions and brainstorming sessions on all aspects of information science and uses of information science and technology to assess and solve institutional, social and research problems. These are informal talks. Discussions are often inspired by real-world problems being faced by the lead discussant.
Labor And Reward In Science: Commentary on Cassidy Sugimoto’s Program on Info...Micah Altman
Cassidy Sugimoto is Associate Professor in the School of Informatics and Computing, Indiana University Bloomington, who researches within the domain of scholarly communication and scientometrics, examining the formal and informal ways in which knowledge producers consume and disseminate scholarship. She presented this talk, entitled Labor And Reward In Science: Do Women Have An Equal Voice In Scholarly Communication? A Brown Bag With Cassidy Sugimoto, as part of the Program on Information Science Brown Bag Series.
Despite progress, gender disparities in science persist. Women remain underrepresented in the scientific workforce and under rewarded for their contributions. This talk will examine multiple layers of gender disparities in science, triangulating data from scientometrics, surveys, and social media to provide a broader perspective on the gendered nature of scientific communication. The extent of gender disparities and the ways in which new media are changing these patterns will be discussed. The talk will end with a discussion of interventions, with a particular focus on the roles of libraries, publishers, and other actors in the scholarly ecosystem..
Utilizing VR and AR in the Library Space:Micah Altman
Matt Bernhardt is a web developer in the MIT libraries and a collaborator in our program. He presented this talk, entitled Reality Bytes - Utilizing VR and AR in The Library Space, as part of Program on Information Science Brown Bag Series.
Terms like "virtual reality" and "augmented reality" have existed for a long time. In recent years, thanks to products like Google Cardboard and games like Pokemon Go, an increasing number of people have gained first-hand experience with these once-exotic technologies. The MIT Libraries are no exception to this trend. The Program on Information Science has conducted enough experimentation that we would like to share what we have learned, and solicit ideas for further investigation.
For slides and comments see: http://informatics.mit.edu/blog
Creative Data Literacy: Bridging the Gap Between Data-Haves and Have-NotsMicah Altman
The document discusses the benefits of exercise for mental health. Regular physical activity can help reduce anxiety and depression and improve mood and cognitive function. Exercise causes chemical changes in the brain that may help protect against mental illness and improve symptoms.
SOLARSPELL: THE SOLAR POWERED EDUCATIONAL LEARNING LIBRARY - EXPERIENTIAL LEA...Micah Altman
Access to high-quality, relevant information is absolutely foundational for a quality education. Yet, so many schools across the developing world lack fundamental resources, like textbooks, libraries, electricity and Internet connectivity. The SolarSPELL (Solar Powered Educational Learning Library) is designed specifically to address these infrastructural challenges, by bringing relevant, digital educational content to offline, off-grid locations. SolarSPELL is a portable, ruggedized, solar-powered digital library that broadcasts a webpage with open-access educational content over an offline WiFi hotspot, content that is curated for a particular audience in a specified locality—in this case, for schoolchildren and teachers in remote locations. It is a hands-on, iteratively developed project that has involved undergraduate students in all facets and at every stage of development. This talk will examine the design, development, and deployment of a for-the-field technology that looks simple but has a quite complex background.
Laura Hosman is Assistant Professor at Arizona State University, holding a joint appointment in the School for the Future of Innovation in Society and in The Polytechnic School. Her work is action-oriented and focuses on the role for information and communications technology (ICT) in developing countries. Presently, she focuses on ICT-in-education projects, and brings her passion for experiential learning to the classroom by leading real-world-focused, project-based courses that have seen student-built technology deployed in schools in Haiti, Vanuatu, Micronesia, Samoa, and Tonga.
Information Science Brown Bag talks, hosted by the Program on Information Science, consists of regular discussions and brainstorming sessions on all aspects of information science and uses of information science and technology to assess and solve institutional, social and research problems. These are informal talks. Discussions are often inspired by real-world problems being faced by the lead discussant.
Making Decisions in a World Awash in Data: We’re going to need a different bo...Micah Altman
In his abstract, Scriffignano summarizes as follows:
l explore some of the ways in which the massive availability of data is changing and the types of questions we must ask in the context of making business decisions. Truth be told, nearly all organizations struggle to make sense out of the mounting data already within the enterprise. At the same time, businesses, individuals, and governments continue to try to outpace one another, often in ways that are informed by newly-available data and technology, but just as often using that data and technology in alarmingly inappropriate or incomplete ways. Multiple “solutions” exist to take data that is poorly understood, promising to derive meaning that is often transient at best. A tremendous amount of “dark” innovation continues in the space of fraud and other bad behavior (e.g. cyber crime, cyber terrorism), highlighting that there are very real risks to taking a fast-follower strategy in making sense out of the ever-increasing amount of data available. Tools and technologies can be very helpful or, as Scriffignano puts it, “they can accelerate the speed with which we hit the wall.” Drawing on unstructured, highly dynamic sources of data, fascinating inference can be derived if we ask the right questions (and maybe use a bit of different math!). This session will cover three main themes: The new normal (how the data around us continues to change), how are we reacting (bringing data science into the room), and the path ahead (creating a mindset in the organization that evolves). Ultimately, what we learn is governed as much by the data available as by the questions we ask. This talk, both relevant and occasionally irreverent, will explore some of the new ways data is being used to expose risk and opportunity and the skills we need to take advantage of a world awash in data.
Software Repositories for Research-- An Environmental ScanMicah Altman
This document provides a summary of the state of software curation based on an environmental scan of research software repositories and related practices. The summary finds:
1) There are no comprehensive indices of software archives and orders of magnitude fewer software archives than data archives. Institutional repositories offer little functionality for software archiving.
2) Very few funders have policies addressing software curation. There is little available advice for researchers who wish to curate, cite, and preserve software.
3) Substantial reproducibility failures continue to be reported due to a lack of software preservation. In summary, software curation looks a lot like data curation did a decade ago, with no universal standards for citing and archiving software.
The Open Access Network: Rebecca Kennison’s Talk for the MIT Prorgam on Infor...Micah Altman
Rebecca Kennison, who is the Principal of K|N Consultants, the co-founder of the Open Access Network; and was was the founding director of the Center for Digital Research and Scholarship, gave this talk on Come Together Right Now: An Introduction To The Open Access Network as part of the Program on Information Science Brown Bag Series.
Gary Price, MIT Program on Information ScienceMicah Altman
This document discusses maximizing the use of open web resources in libraries. It argues that libraries should better utilize free and openly available web content for research and users. However, curating and selecting quality resources from the vast amount on the open web presents challenges including the volume of content, lack of metadata, scalability, and ephemeral nature of some resources. The document outlines potential workflows for discovering, ingesting, reviewing, archiving, and sharing open web resources and suggests tools that can help with curation tasks. It also discusses the types of materials that could be curated from the open web like reports, datasets, digital collections, and videos.
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc
How does your privacy program stack up against your peers? What challenges are privacy teams tackling and prioritizing in 2024?
In the fifth annual Global Privacy Benchmarks Survey, we asked over 1,800 global privacy professionals and business executives to share their perspectives on the current state of privacy inside and outside of their organizations. This year’s report focused on emerging areas of importance for privacy and compliance professionals, including considerations and implications of Artificial Intelligence (AI) technologies, building brand trust, and different approaches for achieving higher privacy competence scores.
See how organizational priorities and strategic approaches to data security and privacy are evolving around the globe.
This webinar will review:
- The top 10 privacy insights from the fifth annual Global Privacy Benchmarks Survey
- The top challenges for privacy leaders, practitioners, and organizations in 2024
- Key themes to consider in developing and maintaining your privacy program
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Monitoring and Managing Anomaly Detection on OpenShift.pdfTosin Akinosho
Monitoring and Managing Anomaly Detection on OpenShift
Overview
Dive into the world of anomaly detection on edge devices with our comprehensive hands-on tutorial. This SlideShare presentation will guide you through the entire process, from data collection and model training to edge deployment and real-time monitoring. Perfect for those looking to implement robust anomaly detection systems on resource-constrained IoT/edge devices.
Key Topics Covered
1. Introduction to Anomaly Detection
- Understand the fundamentals of anomaly detection and its importance in identifying unusual behavior or failures in systems.
2. Understanding Edge (IoT)
- Learn about edge computing and IoT, and how they enable real-time data processing and decision-making at the source.
3. What is ArgoCD?
- Discover ArgoCD, a declarative, GitOps continuous delivery tool for Kubernetes, and its role in deploying applications on edge devices.
4. Deployment Using ArgoCD for Edge Devices
- Step-by-step guide on deploying anomaly detection models on edge devices using ArgoCD.
5. Introduction to Apache Kafka and S3
- Explore Apache Kafka for real-time data streaming and Amazon S3 for scalable storage solutions.
6. Viewing Kafka Messages in the Data Lake
- Learn how to view and analyze Kafka messages stored in a data lake for better insights.
7. What is Prometheus?
- Get to know Prometheus, an open-source monitoring and alerting toolkit, and its application in monitoring edge devices.
8. Monitoring Application Metrics with Prometheus
- Detailed instructions on setting up Prometheus to monitor the performance and health of your anomaly detection system.
9. What is Camel K?
- Introduction to Camel K, a lightweight integration framework built on Apache Camel, designed for Kubernetes.
10. Configuring Camel K Integrations for Data Pipelines
- Learn how to configure Camel K for seamless data pipeline integrations in your anomaly detection workflow.
11. What is a Jupyter Notebook?
- Overview of Jupyter Notebooks, an open-source web application for creating and sharing documents with live code, equations, visualizations, and narrative text.
12. Jupyter Notebooks with Code Examples
- Hands-on examples and code snippets in Jupyter Notebooks to help you implement and test anomaly detection models.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
Ivanti’s Patch Tuesday breakdown goes beyond patching your applications and brings you the intelligence and guidance needed to prioritize where to focus your attention first. Catch early analysis on our Ivanti blog, then join industry expert Chris Goettl for the Patch Tuesday Webinar Event. There we’ll do a deep dive into each of the bulletins and give guidance on the risks associated with the newly-identified vulnerabilities.
GraphRAG for Life Science to increase LLM accuracyTomaz Bratanic
GraphRAG for life science domain, where you retriever information from biomedical knowledge graphs using LLMs to increase the accuracy and performance of generated answers
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slackshyamraj55
Discover the seamless integration of RPA (Robotic Process Automation), COMPOSER, and APM with AWS IDP enhanced with Slack notifications. Explore how these technologies converge to streamline workflows, optimize performance, and ensure secure access, all while leveraging the power of AWS IDP and real-time communication via Slack notifications.
Fueling AI with Great Data with Airbyte WebinarZilliz
This talk will focus on how to collect data from a variety of sources, leveraging this data for RAG and other GenAI use cases, and finally charting your course to productionalization.
Webinar: Designing a schema for a Data WarehouseFederico Razzoli
Are you new to data warehouses (DWH)? Do you need to check whether your data warehouse follows the best practices for a good design? In both cases, this webinar is for you.
A data warehouse is a central relational database that contains all measurements about a business or an organisation. This data comes from a variety of heterogeneous data sources, which includes databases of any type that back the applications used by the company, data files exported by some applications, or APIs provided by internal or external services.
But designing a data warehouse correctly is a hard task, which requires gathering information about the business processes that need to be analysed in the first place. These processes must be translated into so-called star schemas, which means, denormalised databases where each table represents a dimension or facts.
We will discuss these topics:
- How to gather information about a business;
- Understanding dictionaries and how to identify business entities;
- Dimensions and facts;
- Setting a table granularity;
- Types of facts;
- Types of dimensions;
- Snowflakes and how to avoid them;
- Expanding existing dimensions and facts.
Main news related to the CCS TSI 2023 (2023/1695)Jakub Marek
An English 🇬🇧 translation of a presentation to the speech I gave about the main changes brought by CCS TSI 2023 at the biggest Czech conference on Communications and signalling systems on Railways, which was held in Clarion Hotel Olomouc from 7th to 9th November 2023 (konferenceszt.cz). Attended by around 500 participants and 200 on-line followers.
The original Czech 🇨🇿 version of the presentation can be found here: https://www.slideshare.net/slideshow/hlavni-novinky-souvisejici-s-ccs-tsi-2023-2023-1695/269688092 .
The videorecording (in Czech) from the presentation is available here: https://youtu.be/WzjJWm4IyPk?si=SImb06tuXGb30BEH .
How to Interpret Trends in the Kalyan Rajdhani Mix Chart.pdfChart Kalyan
A Mix Chart displays historical data of numbers in a graphical or tabular form. The Kalyan Rajdhani Mix Chart specifically shows the results of a sequence of numbers over different periods.
2. Summary (Background)
• Researchers and the institute have an
increasing interest in measuring outputs...
– citations, impact, researcher identifiers,
collaboration, bibliometrics, altmetrics,
scientometrics
• New databases/api’s, measures and tools are
being introduced frequently
• The Library is engaged in R&D in this area
• The Library is engaged with ORCID identifiers
Overview of Citation Analysis
3. Meta-Resources
• Libguides are available for patrons:
– Asking about additional data/API’s
– Asking about researcher identifiers
• IAP Bibliometrics Course materials available
– Include glossaries, bibliographies, examples, tools
• IAP Bibliometrics Course registration for July
– Includes demonstrations, 1-on-1 consulting
• Research program has expertise
– Available for consultation with library liaisons
Overview of Citation Analysis
4. MIT Library Guides
• Author identifiers guide:
libguides.mit.edu/content.php?pid=573578&sid=4729602
• Api’s for scholarly resources:
libguides.mit.edu/apis
• Introduction to impact factors:
libraries.mit.edu/scholarly/publishing/impact-factors/
Overview of Citation Analysis
5. IAP Course Materials
On informatics site:
informatics.mit.edu
Contains:
- More tools, more data
- Worked examples
- Recommended readings
- Glossary of measures
- Directory of tools,
databases
Overview of Citation Analysis
6. IAP Bibliometrics Course
Not Just in January
- Iapril
- JulyAP
(register through the libraries)
Features:
- Worked examples
- Demonstrations
- One-on-one consultation followups
Overview of Citation Analysis
7. Related MIT Libraries Research
• K. Smith-Yoshimura, et al., 2014, Registering Researchers in
Authority Files, OCLC Research.
• Liz Allen, Jo Scott, Amy Brand, Marjorie M.K. Hlava, Micah
Altman, Beyond authorship: recognising the contributions to
research; Nature.
• Data Synthesis Task Group. 2014. Joint Principles for Data
Citation.
• CODATA Data Citation Task Group, 2013. Out of Cite, Out of
Mind: The Current State of Practice, Policy and Technology for
Data Citation. Data Science Journal. 2013;12:1–75.
Slides and reprints available from:
informatics.mit.edu
Overview of Citation Analysis
8. Consultation available to Library Staff
• New measures, complex analyses:
contact research program escience@mit.edu
• Access to/working with MIT internal data:
contact Program Manager for Scholarly
Repository: sthomas@mit.edu
Overview of Citation Analysis