Air Miles Calculator logo

How far is Windsor Locks, CT, from Craig Cove?

The distance between Craig Cove (Craig Cove Airport) and Windsor Locks (Bradley International Airport) is 8466 miles / 13625 kilometers / 7357 nautical miles.

Craig Cove Airport – Bradley International Airport

Distance arrow
8466
Miles
Distance arrow
13625
Kilometers
Distance arrow
7357
Nautical miles
Flight time duration
16 h 31 min
CO2 emission
1 067 kg

Search flights

Distance from Craig Cove to Windsor Locks

There are several ways to calculate the distance from Craig Cove to Windsor Locks. Here are two standard methods:

Vincenty's formula (applied above)
  • 8466.478 miles
  • 13625.475 kilometers
  • 7357.168 nautical miles

Vincenty's formula calculates the distance between latitude/longitude points on the earth's surface using an ellipsoidal model of the planet.

Haversine formula
  • 8465.625 miles
  • 13624.103 kilometers
  • 7356.427 nautical miles

The haversine formula calculates the distance between latitude/longitude points assuming a spherical earth (great-circle distance – the shortest distance between two points).

How long does it take to fly from Craig Cove to Windsor Locks?

The estimated flight time from Craig Cove Airport to Bradley International Airport is 16 hours and 31 minutes.

Flight carbon footprint between Craig Cove Airport (CCV) and Bradley International Airport (BDL)

On average, flying from Craig Cove to Windsor Locks generates about 1 067 kg of CO2 per passenger, and 1 067 kilograms equals 2 353 pounds (lbs). The figures are estimates and include only the CO2 generated by burning jet fuel.

Map of flight path from Craig Cove to Windsor Locks

See the map of the shortest flight path between Craig Cove Airport (CCV) and Bradley International Airport (BDL).

Airport information

Origin Craig Cove Airport
City: Craig Cove
Country: Vanuatu Flag of Vanuatu
IATA Code: CCV
ICAO Code: NVSF
Coordinates: 16°15′53″S, 167°55′26″E
Destination Bradley International Airport
City: Windsor Locks, CT
Country: United States Flag of United States
IATA Code: BDL
ICAO Code: KBDL
Coordinates: 41°56′20″N, 72°40′59″W