Get AI summaries of any video or article — Sign up free
Literature Search Process || Hindi thumbnail

Literature Search Process || Hindi

5 min read

Based on eSupport for Research's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

Start by defining the research question and topic scope before searching, because it drives keyword selection and later screening.

Briefing

A systematic literature search starts with a clear research question and ends with repeatable search-and-refine cycles across multiple databases—so researchers can find what’s already known, locate gaps, and build a defensible foundation for a literature review or systematic review. The process matters because it helps justify a study, identify relevant theory and methods, and surface useful documentation and resources before writing a review paper.

The workflow begins by defining the topic and converting it into a research question. From there, researchers brainstorm keywords and synonyms tied to the subject area, because search engines and academic databases rely heavily on exact or closely related terms. The transcript emphasizes narrowing broad areas into workable search concepts. For example, “brain-computer interface” is treated as a broad domain that can be narrowed into specific dimensions such as EEG-based brain interfaces or motor imagery–based brain-computer interfaces. This narrowing step is presented as the difference between getting vague results and producing targeted, relevant literature.

Next comes constructing a search strategy tailored to the database being used. Rather than using one generic query everywhere, the approach is to adapt keyword combinations and search logic to match each platform. The transcript lists common options for running searches, including Google Scholar, Semantic Scholar, Scopus, Web of Science, and Lens.org—each offering different coverage and indexing. Researchers are encouraged to run searches, collect results, and then repeat with refined queries when needed.

The middle of the process focuses on why literature searches are performed: to discover existing knowledge on a topic, locate research gaps, justify the need for a study, and identify important theories, ideas, and methodologies. It also highlights practical screening steps such as inclusion and exclusion criteria, applying limits and filters, and using tools like wildcards (for matching word variations) to improve retrieval.

After searching, the workflow shifts to managing outputs. Results should be documented and organized so they can be used later in writing. The transcript mentions saving searches and notes, exporting references (including formats like BibTeX), and using reference management tools such as Zotero and Mendeley. The overall aim is to keep the process traceable and efficient—especially when moving toward systematic literature review formats like systematic reviews, meta-analysis, rapid reviews, scoping reviews, mixed-methods reviews, and related review types.

Finally, the search is treated as iterative: run the search, check results, refine keywords and queries, and repeat across databases until the literature set is sufficiently relevant. That loop—define question, build keyword strategy, search across databases, screen with criteria, and manage references—forms the backbone of a systematic literature review pipeline.

Cornell Notes

A strong literature search is built around a defined research question, then executed through iterative keyword strategy and database-specific searching. Researchers brainstorm keywords and synonyms, narrow broad topics into specific sub-areas (e.g., EEG-based or motor imagery–based brain-computer interfaces), and tailor queries to databases such as Google Scholar, Semantic Scholar, Scopus, Web of Science, and Lens.org. The search then becomes repeatable: run, check results, refine queries, and re-run as needed. Inclusion/exclusion criteria, limits/filters, and techniques like wildcards help improve relevance. Finally, results are documented and exported for later writing using tools like Zotero or Mendeley, supporting literature reviews and systematic review workflows.

Why does a literature search need a research question before any database searching begins?

The process starts by defining the topic and converting it into a research question, because the question determines what counts as relevant evidence. Once the question is set, researchers can brainstorm the right keywords and synonyms, decide what dimensions to narrow into, and later apply inclusion/exclusion criteria consistently. Without that anchor, searches tend to return broad, unfocused results that are harder to screen and justify in a literature review.

How does keyword narrowing improve search results in academic databases?

Keyword narrowing turns a broad domain into specific, searchable sub-concepts. The transcript’s example treats “brain-computer interface” as broad, then narrows it into EEG-based brain interfaces and motor imagery–based brain-computer interfaces. This narrowing helps produce more targeted results because queries match the terminology used in papers, and it clarifies which application dimensions the literature should cover.

What does “construct a search strategy” mean in practice?

It means tailoring queries to each database rather than using one simple search everywhere. The transcript describes building keyword combinations (including synonyms) and adapting them to the database’s search behavior. It also recommends using techniques like wildcards to capture word variations, then applying limits and filters to reduce noise. The strategy is validated by checking results and refining the query when needed.

Why run searches across multiple databases instead of relying on a single platform?

Different databases index different journals, conferences, and records, so coverage varies. The transcript encourages using several sources—such as Google Scholar, Semantic Scholar, Scopus, Web of Science, and Lens.org—to broaden retrieval. It also suggests repeating the search with refined queries after reviewing what each database returns, improving the completeness of the literature set.

How do inclusion/exclusion criteria and screening steps fit into the search process?

After collecting search results, researchers apply inclusion and exclusion criteria to decide which papers belong in the review set. The transcript also mentions using limits and filters and selecting relevant databases to keep results aligned with the research question. This screening step is essential for producing a defensible literature review, especially for systematic or scoping reviews.

What should happen to search results after retrieval so writing becomes easier?

Results should be documented and managed for later use. The transcript highlights saving searches and notes, exporting references (including BibTeX), and using reference management tools like Zotero or Mendeley. This turns the search output into a reusable evidence library for writing the literature review and for subsequent review types such as systematic reviews, meta-analysis, and scoping reviews.

Review Questions

  1. What steps convert a broad topic into a searchable research question and keyword set?
  2. How would you narrow a broad domain (like “brain-computer interface”) into multiple sub-areas for better retrieval?
  3. Why is it necessary to refine queries and repeat searches across databases rather than running a single search once?

Key Points

  1. 1

    Start by defining the research question and topic scope before searching, because it drives keyword selection and later screening.

  2. 2

    Brainstorm keywords and synonyms, then narrow broad concepts into specific dimensions to improve relevance.

  3. 3

    Build database-specific search strategies instead of using one generic query across platforms.

  4. 4

    Run searches across multiple databases (e.g., Google Scholar, Semantic Scholar, Scopus, Web of Science, Lens.org) to improve coverage.

  5. 5

    Use inclusion/exclusion criteria, limits, filters, and techniques like wildcards to reduce irrelevant results.

  6. 6

    Treat searching as iterative: run, review results, refine queries, and repeat until the set is sufficiently relevant.

  7. 7

    Export and manage references using tools like Zotero or Mendeley (including formats like BibTeX) to support later writing and review workflows.

Highlights

A systematic search hinges on a research question that guides keyword brainstorming, narrowing, and screening.
Narrowing broad terms (e.g., brain-computer interface) into specific sub-areas (EEG-based vs motor imagery–based) improves retrieval quality.
Search strategy should be tailored to each database, then refined through repeated run-and-check cycles.
Inclusion/exclusion criteria and filters are essential for turning raw search results into a usable review set.
Reference management tools like Zotero or Mendeley help preserve search outputs for later literature review writing.

Topics

Mentioned