Does everyone need to take vitamins — and which are most important? Here’s what to know
More than half of American adults took dietary supplements between 2017 and 2020 — but are vitamins really essential for everyone? Experts share what you need to know.