Horizontal Projection
1. **State the problem:** A ball is projected horizontally from a height of 20 m with an initial horizontal velocity of 10 m/s. We need to find how far from the base of the projection point the ball will land.
2. **Relevant formulas:**
- The time of flight for a horizontally projected object is determined by the vertical motion under gravity: $$t = \sqrt{\frac{2h}{g}}$$ where $h$ is the height and $g$ is the acceleration due to gravity (approximately 9.8 m/s²).
- The horizontal distance traveled is given by: $$d = v_x \times t$$ where $v_x$ is the horizontal velocity.
3. **Calculate the time of flight:**
$$t = \sqrt{\frac{2 \times 20}{9.8}} = \sqrt{\frac{40}{9.8}} \approx \sqrt{4.08} \approx 2.02 \text{ seconds}$$
4. **Calculate the horizontal distance:**
$$d = 10 \times 2.02 = 20.2 \text{ meters}$$
5. **Interpretation:** The ball will land approximately 20 meters from the base of the projection point.
**Final answer:** 20 m (option b)