Communications - Mission Viejo City Council

Transcription

Communications - Mission Viejo City Council
ATS
Communications
Managing the Wireless Net
unicipal Agencies to Monetize
The 5 Reasons For
1.
2.
3.
4.
5.
To avoid the risk of lease termination
To fund capital improvements
To fund investment opportunities
To reduce debt
To obtain capital for operating costs
Risk Factors
Mergers AT&T/Cingular and Sprint/Nextel
New Technology WiFIJWiMax, DAS networks, Wireless mesh, Femtocells
Changes in Government Regulation
Operating Efficiency —Decommissioning of an old site may cost less than to upgrade to a new
site.
e) Rent reduction currently there are two companies that are actively working for the carriers
to reduce lease rents
a)
b)
c)
d)
—
—
—
CALIFORNIA OFFICE
22264 Lambert St., Suite 402, Lake Forest, CA 92630, Phone: (949) 305-7848 Fax: (949) 768-6984
Website www.atscomm.com + email: [email protected]
Table of Contents
Supporting Reference for Risk Factors
1.
ergers
2. New Technology
3. Changes in Government Regulations
Supporting Reference for Monetization
4. To Fund Investment Opportunities
5. To Fund Capital Improvements
Managing the Wireless Net
Supporting References for Risk Factors
Mergers
Consolidation is one of the greatest threats to continuing a landlord’s lease income. Mergers have effects
on the companies that join forces, some mergers happen smoothly and the newly acquired company fits
very nicely in the portfolio, while other companies have a hard time getting back on their feet after a large
number of internal changes occur. For example, many assets are acquired during a merger that add to
operating costs of expenses. As you will read in the articles below, “cell site leases represent the
first or second largest expense for wireless carriers internationally”. When two wireless
carriers merge many times their sites and coverage areas overlap, the high cost of operation incentivizes
the carrier to get the unnecessary sites off air in order to reduce their expenses.
“American Tower said an AT&T Wireless- Cingular merger would mean overlap on 350 of
its sites, resulting in an overlap of up to $9 million in annual revenue...”
“SpectraSite said that in a “worst-case churn scenario,” it would lose 450 leases, or 3 percent
of its portfolio.., due to an AT&T Wireless -Cingular merger.”
Source: RCR Wireless News, January 26, 2004
Why Carriers are Retaining Real Estate Asset Managers
“It’s time for American corporations to take a closer look at their real estate portfolios.”
Commercial Investment Real Estate
A reevaluation of the way corporations handle their real estate is especially important in
light of the anticipated growth in buying and selling, says a recent Coldwell Banker
Commercial/National Real Estate Investor corporate real estate survey reported in CIRE.
This forecast is particularly significant to the wireless communications industry. With
nearly 184,000 cell sites nationwide (which reflects a growth trend of 4.5% year over
year since 1985), wireless carriers have officially acknowledged the magnitude of their
real estate portfolios. In fact, cell site leases represent the first or second largest
exnense for wireless carriers internationally. Managing these costs has become a
number-one priority as carriers navigate their forward vlannin and the new technology
that will nrofoundlv impact the growth of their cell site networks.
A by-product of the wireless industry’s growth strategy since its inception, carriers’ real
estate portfolios have been a largely overlooked asset for 20 years. Now, the industry is
taking steps to manage their commercial real estate interests with the same savvy
indicative of corporations focused on maximizing their investments.
But commercial real estate is a complex business and building owners for decades have
tapped industry experts to represent their interests. Professional landlords or
partnerships that own small office buildings often employ their own brokerage listing
agents who monitor the competition’s lease terms and conditions, while large office
complexes owned by institutional investors employ professional property and asset
managers.
An estimated three to ten percent of brokers focus exclusively on representing tenants,
a more recent industry segmentation that equalizes access to critical market
information. Says CIRE magazine, most Fortune 1000 corporations utilize tenant rep
services.
“When tenants do not go to the marketplace to fully determine whether the renewal
terms and conditions are competitive, they almost always lose. Knowledgeable landlords
keep pace with the marketplace and fully appreciate the financial rewards to the
property owner when tenants enter lease renewal negotiations without a broker and
without bringing alternative locations into competitive play.”
This commercial real estate management strategy is overtaking the wireless industry
and a new niche has developed for the administration and evaluation of cell site leases
real estate asset managers for the wireless industry. With their sprawling nationwide
network that functions in dozens of markets simultaneously, carriers require yet
another level of specialized tenant representation.
—
© Copyright 2006
http:I/www.md7.comlrgp_articles.html
RCRXVi1e1ess News
Sprint Nextel moves into junk status, gets new CFO
By Kristen Beckman
Story posted: May 2, 2008 1:28 pm EDT
-
Standard & Poor’s Ratings Services lowered its corporate credit and senior unsecured ratings on Sprint Nextel Corp.
to BB from BBB- and removed the company’s ratings from CreditWatch with negative implications. The downgrade
moves Sprint Nextel out of the investment-grade category and into junk status.
“The downgrade is based on our assessment that Sprint Nextel’s business risk profile is no longer supportive of an
investment-grade rating given its deteriorating operating performance and lack of visibility in the wireless business,
along with increased financial leverage due largely to declining EBITDA,” said Standard & Poor’s credit analyst Allyn
Arden
The ratings firm placed Sprint Nextel on CreditWatch in February after the company warned of steep postpaid
subscriber losses. S&P said it expects Sprint Nextel’s operating and financial results to remain under pressure over
the next couple of years.
The downgrade coincided with the arrival of a new CFO for Sprint Nextel. Robert H. Brust was named its new CFO
effective immediately. Brust previously was CFO at Eastman Kodak Co. and CFO of Unisys Corp. He also spent 31
years with General Electric.
William Arendt, senior VP and controller, has been serving as interim CFO since Paul Saleh and two other executives
were ousted earlier this year in a management shakeup.
The company is expected to report first-quarter results May 12.
PRINTED FROM: http:/lwww.rcrwirelessnews.com/apps/pbcs.dII/article?AID=/200805021FREE161 994528/1 OO2IrssOl &template=printart
New Technology
No one really knows what’s going to be happening 10 to 20 years from now. We have seen leaps in
technology within the past 20 years, which is when the Internet came into play. Going back vely few
would have been able to conceive of cell phones that would run our personal or business activities.
Recently, due primarily to cost, there have been new advances in new wireless infrastructure deployment.
Femtocells and open DAS networks will affect traditional macro-cell deployment. We have also seen
other technologies that have become obsolete due to new technological developments, such as analog and
frequency spectrums that have been re-allocated to the wireless industry (formerly broadcast television).
The development and deployment of signal combining technologies, which permit one antenna to sen/ice
multiple frequencies and, thereby, multiple customers, may reduce the need for our antenna space. In
addition, other technologies which may be developed and emerge may serve as substitutes and
alternatives to leasing which might otherwise be anticipated or expected on our sites had such
technologies not existed.
Source: RCR Wireless News, January 26, 2004
NE OR
‘RID
This story appeared on Network World at
http://www.networkworld.com/su pp/2006/ndc6/102306-ndc-wireless-mesh.html
Wireless mesh or WIMAX: Which MAN
technology is best?
Enterprises give wireless mesh a whirl
--
and they like the ride.
By Joanne Cummings, Network World, 10/23/2006
Not too long ago, enterprises had limited options when it came to making a wireless network part of the New Data
Center plan. In the local area, the choice was Wi-Fl 802.llb, g or a. In the larger campus or metropolitan area,
people anticipated widespread WIMAX deployments (aka 802.16d or 802.16e) because of the promise of
broadband-level bandwidth, improved flexibility and tight security.
-
Not so anymore. These days, metropolitan wireless mesh networks threaten to leave WIMAX at the gate. The
reasons are numerous: They follow the traditional Wi-Fi standard, so any Wi-Fl-enabled client can work with them,
users and analysts say. Plus, they don’t require wire runs to every node, they are designed to be self-organizing
and -healing, and they scale on the fly. If you need more capacity, add another node, and you’ve got it.
In addition, mesh gear is available from Cisco, Firetide, Strix Systems, Tropos Networks and others. WIMAX
products are not nearly as prevalent. Fixed WiMAX equipment is just coming to market, and mobile WIMAX gear is
not expected to arrive much before late 2007.
“I looked at WiMAX in the public safety arena.... It worked great in a stationary environment, especially for video
cameras, but once I tried it in a mobile environment, it didn’t work,” says Peter Collins, dO for the city of Austin,
Texas, which recently deployed a series of wireless meshes using Cisco Aironet 1500 series gear.
That’s not to say mesh is always the best choice. For most companies considering deploying in-building wireless
networks, the cost equation indicates traditional Wi-Fi is still the way to go if the wire is there and available for
the Wi-Fi access points. But when the infrastructure has not been wired, and running wire would be prohibitively
difficult or expensive, mesh makes sense. “Mesh works well, and we’re starting to see it in warehouses, loading
docks, logistics, transportation, those kinds of applications. We also see it in large, temporary setups, like outdoor
concert venues,” says Craig Mathias, principal at the Farpoint Group.
-
For example, North American Midway Entertainment, a large amusement company in Los Angeles, sets up Firetide
wireless mesh networks to support transaction processing for its fairs and carnivals. Firetide’s HotPort mesh
network gear provides reliable, robust connectivity in a challenging environment, says John Gallant, North
American Midway’s CIO.
“For us, it’s almost impossible to use a conventional wireless network with point-to-point runs and wireless access
points,” he says. “The Firetide mesh gives us a multipoint setup. We can actually go around corners. If the
connection’s strongest point is straight in front of you, but every three minutes a huge mechanical device like a
ride goes past it for a minute, that creates a lot of delay and latency. With the Firetide gear, if one node becomes
obstructed or the network noise level gets too high, the node will automatically route the signal to the next best
possible route,” he adds.
In mesh networks, users deploy multiple mesh nodes throughout an area, but, unlike in traditional Wi-Fi, only one
node has to be connected to a wired network. When a mesh node receives a frame from a Wi-Fi client, it relays the
frame from node to node, until the frame reaches the wired node. Each node has a standard Wi-Fi interface, to
communicate with clients, and a radio-based backbone link that relays the message across the network.
Because the mesh doesn’t require wire runs to every node, Gallant can use it in the large spaces he needs to cover.
For example, North American Midway recently ran the Canadian National Exhibition in Toronto, where the
fairgrounds blanketed about one square mile. Gallant implemented a mesh comprising 42 nodes and access points.
It supported 240 users.
Mesh has its downsides, he says, but those are common to all wireless networks susceptibility to lightning,
interference from other devices, power outages. Most problematic is having to plug nodes into standard, 110-volt
household receptacles. “They have battery backups in case you lose power, but you do have to plug them in. That’s
probably the biggest challenge,” Gallant says. Still, reliability outweighs the downsides: “Mesh works great,” he
says.
-
Proprietary routing standards
A bigger downside for some users considering mesh networks is the proprietary nature of each vendor’s routing
scheme. That means once you make the decision to buy, you’re locked into that vendor. Although the IEEE is
working on a standard routing protocol, called 802.lls, it doesn’t expect to finish that work until late 2007. Even
then the standard would provide for fairly vanilla multivendor mesh implementations.
This gives some users pause. “Wireless mesh is a little bit out there,” says Elliot Zeltzer, global manager for
telecommunications
esh vs. i AX
security at General Motors
in Detroit. “We can’t afford
any
downtime.
We’re
Wireless mesh (Wi-Fl)
WiMAX
more
looking
for
standardized
established,
Bandwidth
11M or 54Mbps
70M
100Mbps
technology.”
.
.
.
.
-
to
-
Coverage
No limit
No limit
GM is in the middle of a
traditional Wi-Fi rollout
Availability
Nowt
By year-end’2OO
across its campus, primarily
because the wiring is there
and it’s a more proven technology, Zeltzer says. If he were to look at a more metropolitan-level rollout, he says
he’d favor WIMAX.
.
-
-
“Not speaking specifically to GM, I believe WiMAX has the highest value,” Zeltzer says. “It offers huge amounts of
bandwidth for fairly low cost and in the end, it could provide total local-carrier bypass,” he says. Plus, Zeltzer says,
he is leery of mesh security.
Wireless mesh users counter such arguments by saying mesh offers enough coverage, bandwidth and security for
today’s applications. “Everything’s misleading in the world of wireless, and your rate depends on a lot of factors,
like distance and interference,” Austin’s Collins says. “With W1MAX, your rate decreases the farther away you are
from the transmitter and receiver. It’s all relative. I have a consistent throughput on wireless mesh, and that’s
more important.”
As for security, Farpoint’s Mathias, as well as current users, say it’s not an issue: “Basic wireless LAN security is
really improved to the point where it’s very good,”he says. “If you can secure a wired network, you can secure a
wireless network.”
In the end, the proof is in the deployments. Several cities have committed to wireless mesh rollouts, and vendors
seem to be ticking off new users weekly. As Mathias says: “The demand. is a global phenomenon.”
.
.
Cummings is afreelance writer in North Andover, Mass. She can be reached [email protected].
All contents copyright 1995-2008 Network World, Inc. http://www.networkworld.com
Femtocells: Spectrum Efficiency Improved
An article by Manish Singh, VP Product Line Management, Continuous Computing
From the May 12, 2008 edition of Wireless Design & Development
Since wireless spectrum is at a premium, it is advantageous
for wireless carriers to use the spectrum efficiently in both
outdoor and indoor environments.
One of the most limited resources in telecom is wireless spectrum; therefore, it is not surprising
that carriers are paying premium licensing fees for spectrum acquisition. For example, the recent
700 Mhz spectrum auction in the U.S. had an aggregate bid value of $19.6 billion. Since
acquisition of wireless spectrum is limited, it is advantageous for wireless carriers to use the
spectrum efficiently in both outdoor and indoor enviromnents.
From a radio frequency (RF) propagation perspective, buildings and walls have never been
friends to a radio signal. Physical structures cause RF signal attenuation because of reflection,
diffraction, scattering, and multi-path signal fading, all resulting in poor radio signal reception in
an indoor environment. In some cases, virtual
dead spots can occur inside buildings, even
3G Fomtocoll Solution
though there is great coverage outside. Walls
can attenuate signals by 10 to 20 dB depending
•Oflloadst omost
• Providos cc coverage ndoors
on the type of construction. At the same time,
power-hungry
• Low-power and soit-contigu ng to
3G users trom the
m nm}zo bntororonco
studies have shown that an average consumer
macroco to home
• Condensed mob o network (RAN)
spends 50 to 60% of his or her time in an indoor
base so uflon
• Leverages customo broadband to
Ca ry traffic to operators network s rig
environment, and that 70% of wireless calls
ub SIP and UMA
originate and terminate indoors. All things j
considered, this poses the question: Are wireless
carriers using their spectrum efficiently in
indoor environments?
Market Drivers for Femtocells
From an engineering standpoint, wireless
system spectral efficiency is measured in
bps/Hz/cell. From a carrier standpoint, this
Macro Network
Standard
3G localized
translates into the number of simultaneous
Base Station
air Interface
3G mobile phone
subscribers that can be supported in one cell site
(i.e., network capacity). From a consumer
standpoint, this all translates into quality of
voice and video, dropped calls, and high-speed packet access (HSPA) data rates. Poor indoor RF
signal penetration has adverse effects on the link spectral efficiency, which has a direct impact
on system spectral efficiency. Simply put, carriers lose revenue as mobile minutes are lost to
landline minutes when consumers experience dropped calls, poor quality of voice, and lower
data rates when indoors. When considered in aggregate, one can easily see the underlying market
drivers for femtocells.
What is Femtocell?
A femtocell is essentially a small wireless base station that resides in the consumer’s home or
office. Femtocells transmit at very low power, yet create almost ideal indoor radio conditions.
For backhaul, femtocells use an IP broadband connection (e.g., fiberfDSL/cable). The very walls
which are a radio signal’s adversary actually become their friend as they attenuate RF signal
propagation out of the home from the femtocell, thereby minimizing radio interference with an
existing macro-cellular network or another nearby femtocell. In fact, femtocells not only provide
excellent indoor coverage, but also, by virtue of creating a small indoor cell-site, they free up
capacity in the macro-cellular network. In other words, indoor subscribers’ cell phone traffic is
parked on a femtocell as opposed to on a macrocell. Femtocells eliminate the need for dual-mode
handsets as virtually any existing wireless handset should seamlessly work with a femtocell. And
last but not least, one femtocell can support four to six simultaneous voice calls, which means
that each member of a family of four can talk simultaneously as if they had four virtual landlines
for the price of one.
Why Has it Taken so Long?
So, if femtocells enhance coverage, increase network capacity, and improve spectral efficiency,
then why have they not been widely deployed before now? In fact, the femtocell is not a new
idea. It has been around for a long time, yet the catalyst for driving a commercially viable rollout
for femtocells has only fallen into place very recently. In the earlier days, backhaul costs were a
barrier for mass deployment, but today the wide proliferation of IP broadband connections (fiber,
DSL, or cable) means that the backhaul infrastructure already exists. In the U.S. alone, more than
52% of homes now have broadband connections, and the figure is growing. Femtocells leverage
a consumer’s IP broadband connection to backhaul voice, video, SMS, and data traffic from
millions of homes into an existing wireless core network.
Furthermore, thanks to Moore’s Law
which states that silicon performance doubles every 18
we are now at the point where integrated custom silicon parts for femtocells are
months
widely available. One silicon part, integrating the required compute, digital signal processing,
encryption and various other key features for a small wireless base station, is key to drive the bill
of materials (BOM) cost down for a commercially viable femtocell deployment. Lastly, with
more than 2.5 billion wireless subscribers worldwide, the economy of scale to justify femtocell
deployment is finally in place.
—
—
Femtocell in Enterprise
The business case for femtocell rollout is straightforward and can be summarized in three key
parts. First, consumers today are paying for landline connections. These landline minutes and
Average Revenue Per User (ARPU) are up for grabs due to competition, but only if wireless
carriers can provide reliable indoor coverage. In the U.S., the number of cell phone-only homes
exceeded the number of landline-only homes in 2007, and household cell phone spending has
exceeded landline spending. With femtocells, carriers can provide a robust wireless alternative to
landlines and accelerate this social phenomenon of consumers replacing their landlines.
Second, the demand for mobile broadband is on the rise, as fueled by the YouTubes and
Facebooks of the world. In the first half of 2007, data revenue in the U.S. alone was 15.5% of all
wireless revenue. For AT&T, data ARPU now constitutes 18.4% of their total ARPU, and
Vodafone’s data revenue grew 49% over the past year. Femtocells add network capacity and
make it possible to deliver 7.2 Mb/s and 14.4 Mb/s HSDPA data rates to consumers in indoor
thereby fulfilling market demand for wireless broadband access.
environments
—
Thirdly, femtocells deliver capital and operational expenditure savings. Adding a new macro
base station costs roughly $600,000, depending on site geography, technology, supplier cost
model, etc. Add to this approximately $15,000 per month for operation costs, such as leased
lines, electricity, and cooling, etc. For the cost of one macrocell site, a network operator could
potentially cover 6,000 homes, assuming a cost of $150 femtocell price per unit and operator
subsidies of $100 per unit. For $600,000, assuming 7.2 Mb/s HSDPA rate, one macrocell site
would add an aggregated 7.2 Mb/s bandwidth capacity, as opposed to 43 Gb/s with 6,000
especially if revenue is accumulated on a usage basis.
femtocells. This is a huge difference
—
Carrier Benefits
The carriers that roll out femtocells first will benefit from reduced churn as subscribers will not
struggle with dropped calls. The quality of the consumer’s experience will improve significantly
as clear voice and HSPA data rates are delivered. Most importantly, the carriers that initiate the
charge will benefit from “reverse-churn” as femtocells enable home-zone “family tariff plans”
which will entice all family members to switch to one carrier in order to take advantage of
packaged bulk service benefits.
Second, the demand for mobile broadband is on the rise, as fueled by the YouTubes and
Facebooks of the world. In the first half of 2007, data revenue in the U.S. alone was 15.5% of all
wireless revenue. For AT&T, data ARPU now constitutes 18.4% of their total ARPU, and
Vodafone’s data revenue grew 49% over the past year. Femtocells add network capacity and
make it possible to deliver 7.2 Mb/s and 14.4 Mb/s HSDPA data rates to consumers in indoor
thereby fulfilling market demand for wireless broadband access.
environments
—
Challenges of Femtocell Rollout
While the economics for femtocells are compelling, there are many technical challenges that
need to be addressed. Utilizing femtocell means that a wireless base station essentially moves
into the consumer’s home, and along with that comes the challenge of integrating millions of
femtocells with an existing 3G wireless core infrastructure. Many different architecture options,
like Tunneled lub, Tunneled hi, and SIP/IMS are under consideration, and much more work is
needed in order to reduce these choices to only one or two.
Femtocells also bring up the challenge of managing RF interference. RF interference between
femto-femto and femto-macro can create virtual dead spots as well as the potential of degrading
the performance of an existing carrier’s macro network. Traditional RF planning tools used to
plan macro cellular networks will not work for femtocell deployments. Instead, femtocells will
need the ability to scan the dynamically changing RF environment and configure themselves
with frequency bands, scrambling codes and transmit power levels.
Femtocells take security challenges to another level. Since the traffic from femtocells can be
carried over a public IP network, the traffic needs to be encrypted to avoid eavesdropping.
Potential solutions include establishing IPSec tunnels, SRTP, or TLS for securing the traffic.
Carriers also need to secure their existing core network infrastructure from any Denial of Service
OS) attacks
hence the need for a Security Gateway between femtocells and the core
network.
—
Quality of Service (QoS) is another challenge that needs to be addressed. The best effort of an IP
network is just not good enough to carry real-time voice which might be vying for bandwidth on
a shared IP link with peer-to-peer (P2P) and other data traffic. In certain deployment scenarios,
the DSL uplink bandwidth can be as low as 200 Kb/s and pack four simultaneous voice calls. In
this environment, not losing a consumer’s VPN connection on his or her PC will be difficult. As
a result, classifying, marking, and prioritizing [P packets based on class of service becomes a key
requirement. Potential QoS solutions include static Diffserv-based solutions as well as Signaled
QoS solutions based on RSVP or SIP.
TR-069 Standard
Management of millions of femtocells requires unique solutions. Traditional Element
Management Systems (EMS) that are used to manage NodeBs and Radio Network Controllers
(RNCs) in today’s 3G networks simply do not scale for femtocells. The DSL Forum already has
developed the TR-069 standard for remote customer premise equipment (CPE) device
management. This standard needs to be further extended to support femtocells’ data models and
deployment models. With TR-069, carriers can re-use their existing CPE device management
infrastructure by just extending it to include femtocells. Likewise, to support E9 11 regulations,
carriers will need the exact location information of femtocells to effectively route emergency
calls to the nearest public safety answering point (PSAP). Femtocell location information could
potentially be obtained using global positioning system (GPS) technology; however, GPS itself
does not function very well in indoor environments. Another alternative could be mapping a
static IP address to a femtocell’s physical location. All of these challenges will need to be
addressed before widespread femtocell deployment becomes a reality.
What Lies Ahead?
Looking further into the future, femtocells will be more than just access points. In 2008, the
market will see more converged CPE devices with femtocell functionality embedded into a
residential gateway
an IPTV set top box or a cable/DSL modem. For example, Netgear has
already announced the availability of a Residential Gateway incorporating femtocell technology.
These converged CPE devices will enable carriers to deploy true quad-play services
voice,
—
—
all out of one box. And let’s not forget that 30 Long Term
video, data, and mobility
Evolution (LTE), or “4G,” standardization work is already being influenced architecturally as
eNodeBs are also expected to be deployed as home base stations, i.e., Home eNodeBs (HeNB).
Carriers that have already secured CPE real estate in consumers’ homes will continue to have an
advantage long into the next decade.
—
Conclusion
Carriers and consumers will both benefit by femtocell rollouts. Carriers will accelerate return on
investment (ROl) on spectrum acquisition by efficiently using this scarce and costly resource in
indoor environments, while at the same time monetizing mobile minutes substituting for landline
minutes; thereby, creating pull for new customer acquisitions through family tariff plans and
aggressively growing the data ARPU. Consumers will benefit as they jettison their landlines and
become accustomed to their existing cell phones working seamlessly with femtocells while
enjoying better coverage, HSPA data rates, zero dropped calls and multiple virtual connections.
And this is all just the beginning. In the future, femtocells are poised to play a key role in
delivering converged services via IP Multimedia Subsystem (IMS)
a topic for another day and
another article.
—
About The Author
Manish Singh is vice president, product line management at Continuous Computing. Previously
he served as vice president of field engineering and vice president of engineering. Mr. Singh is
an experienced engineering leader who brings 14 years of experience in telecommunications
product design and development. Prior to Continuous Computing, Mr. Singh held various
engineering management and architect positions at Intel Corporation, Trillium Digital Systems,
and C-DOT (Center for Development of Telematics). Mr. Singh holds the patent as a sole
inventor of “Configurable Cache” memory system. Apart from broad telecom domain
knowledge, Mr. Singh brings specialized expertise in the wireless and VoIP telecom technology
areas. In 1998, he successfully led the development of 20 MSC at C-DOT and since has led
various VoIP and 3G/IMS wireless product development. Mr. Singh received a B.S. degree in
electronics and telecommunication from Shri GSlnstitute of Technology & Science, Indore and a
M.S. degree in computer science from India Institute of Science, Bangalore.
RCRWf l-ss News
Report: AT&T Mobility to sell $100 femtocells
Carrier Inked contract with femto vendor lp.access
By Mike Dano
Story posted: April 24, 2008 1:21 pm EDT
-
According to a report from research and banking firm ThinkPanmure, AT&T Mobility plans to sell up to 7 million
femtocells from ip.access Ltd., a picocell and femtocell infrastructure vendor based in Cambridge, United Kingdom.
According to ThinkPanmure, AT&T signed a contract with the firm for up to $500 million in femtocells over the course
of five years, and will sell the devices for as little as $100 each.
A spokeswoman for AT&T did not comment on the report specifically; however, she said the nation’s largest carrier is
examining the potential benefits of femtocells through lab tests, and plans to conduct a femtocell trial “later this year.”
Representatives from ip.access were not immediately available to comment on ThinkPanmure’s report.
Femtocell technology has caused a notable stir in the wireless industry. FemtoceUs essentially are tiny base stations
that can be sold to individual phone users. Femtocells can be installed in homes or offices, and transmit cellphone
traffic through a high-speed Internet connection. Femtocells work much like VV1-Fi hotspots, except they power
transmissions over a wireless carrier’s licensed spectrum.
Thus, femtocells raise a number of potentially troublesome issues for users and carriers, including the management
of radio interference between a cell tower and a femtocell; radio interference between different femtocells, such as in
an apartment building; and roaming between femtocells.
Nonetheless, femtocells could solve one of the biggest headaches wireless carriers face: indoor coverage. The
installation of a femtocell in a home or office would eliminate the need for a carrier to cover that same space, and
offloads the cost of the equipment onto the user.
And a number of carriers have embraced the technology. Sprint Nextel Corp. was the first out of the femtocell gate in
the United States; the carrier in September introduced its Airave femtocell from Samsung Electronics Co. Ltd. for $50
along with a monthly fee. The product allows users to make unlimited calls over their femtocell connection.
When the Airave launched, it was only available in Denver and Indianapolis. The carrier expanded distribution to
Nashville in October.
T-Mobile USA Inc., though, is taking a slightly different tack. The carrier last year introduced a dual-mode calling
service that allows users with GSMNV1-Fi phones to place calls over their Wi-Fi network for free.
Although AT&T would not comment on its UMA plans, Current Analysis analyst Peter Jarich said that if the carrier
placed UMA software on its iPhone “they’d probably see a great uptake and they could eventually leverage the core
network kit to support femtocell network integrations as well.”
ThinkPanmure said GSM operators will roll out femtocells on a broad scale by next year.
PRINTED FROM: httpilwww.rcrnews.comlappslpbcs.dlllarticle?AID=/200804241FREE189561 9934/Oifrontpage&template=pnntart
s issinfo.ch
July 10, 2006
-
12:42 PM
Mobile phone airship to conquer stratosphere
Airships like this one may one day fill the stratosphere (Stratxx)
A zeppelin will replace all the terrestrial mobile phone antennas in Switzerland
inventor has his way.
-
if a Swiss
Should Kamal Alavi’s project for the high-tech airship take off, the worlds of mobile telephony and
data transmission would be turned on their heads.
Not only would the technology, called High Altitude Platform Systems (Haps), make the current 1,000
earth-bound antennas redundant, it would drastically reduce radiation.
A Swiss of Iranian extraction, Alavi is a former aerospace engineer turned entrepreneur who heads his
own firm, Stratxx. Together with a team of 50 scientists, he is preparing a 2007 test run of the
airship, which he has named the “X station”.
Thanks to a GPS steering system developed by the Swiss Federal Institute of Technology, the 60meter long helium-filled balloon will remain stationary at 21 kilometres above the earth.
A small-unmanned aircraft outfitted with a mobile phone antenna and other devices for transmitting
digital data will be attached to the zeppelin. The X station has been equipped with giant propellers to
help counter the almost constant buffeting from the wind.
Solar panels will supply the energy to propel the airplane and antenna. Underneath will be a platform
containing technical equipment, conceived by Ruag, the large Swiss aerospace concern.
Radiation
“Transmitting on earth causes lots of radiation, because you have to penetrate countless buildings,”
Alavi says, arguing that phone connections are more reliable when transmitted from above because
the signals are unobstructed by manmade or natural objects.
And “spot beam” antennas developed at Lausanne will allow radiation to be adjusted according to
usage, regions with little activity receiving relatively less.
But Switzerland’s largest mobile telephone operator, Swisscom believes not all of the technological
hurdles have been overcome.
“This project cannot replace the present mobile telephone system,” spokesman Sepp Huber told
swissinfo.
The X station would not be limited to forwarding mobile telephone signals, but would also be capable
of handling the radio, television and internet needs of entire nations.
Alavi believes that his project is also economical. He estimates that a Haps airship will cost no more
than SFr4O million ($32 million).
In comparison, a single mobile phone antenna costs about SFr300,000 while a communications
satellite starts at SFr600 million.
Alavi says the X stations are conceived to be low maintenance. In the event of a defect, the aircraft
will be decoupled from the airship and returned to earth, much like a mini-space shuttle.
The project is now in a key phase. Solar cells are being tested at an altitude of 30 kilometres, and final
preparations are underway for the launch of the first airship into the stratosphere. The entire system
should be ready for testing a year from now.
The potential is enormous if Stratxx manages to be the first to fly with this new technology. About 20
Haps would be required to cover Europe alone while Africa would need twice as many.
swissinfo, Etienne Strebel
CONTEXT
The X station will go to an altitude of 21 kilometres nine more than civilian aircraft are permitted. This
height is needed to place the antenna stations above the jet stream where winds are moderate.
—
Satellites are stationed from 500 to 36,000 kilometres above earth.
An X station can service an area of about 1,000 kilometres in diameter. X stations can also exchange data
between themselves.
In addition to the Swiss Federal Institutes of Technology at Zurich and Lausanne and Ruag, participants in
the project include Neuchâtel University.
http://www.stratxx.com/)
URL of this story:http ://www.swissinfo.ch/eng/swissinfo. html?siteSect= 105&sid =6873540
Changes in Government Regulation
The Federal Communications Commission has the strategic goal for, “All Americans to have
affordable access to robust and reliable broadband services.” The support of the FCC in the
deployment of broadband services through WiMax, as well as other telecommunication services
that are coming into the industry, will bring competition to the signals that are transceived from
wireless facilities. Broadband services allow for the convergence of voice, video, and data
services into a single network. Not only does it allow for a single network, but it can be carried
further over distances and it is cheaper to deploy. Currently, multi-mode phones are being
rd
th
generation technologies, such as VOIP and Internet, which will require
equipped with 3
and 4
additional bandwidth.
The FCC’s support of these services will quickly bring these
marketplace.
technologies to the
The strategic goal goes on to state that, “Regulatory policies
must promote technological neutrality, competition, investment, and innovation to ensure that
broadband service providers have sufficient incentive to develop and offer such products and
services.”
Mobile satellite systems and other new technologies may compete with land-based wireless
communications systems, thereby reducing the demand for tower space and other services we provide.
The FCC has granted license applications for several low-earth orbiting satellite systems that are
intended to provide mobile voice or data services. The growth in delivery of video services by direct
broadcast satellites may also adversely affect demand for our antenna space.
Source: RCR Wireless News, January 26, 2004
Order Code RS20993
Updated May 27, 2004
CRS Report for Congress
Received through the CRS Web
Wireless Technology and Spectrum Demand:
Third Generation (3G) and Beyond
Linda K. Moore
Analyst in Telecommunications and Technology Policy
Resources, Science, and Industry Division
Summary
Advances in wireless telecommunications technology are converging with Internet
technology to foster new generations of applications and services. Presently, the United
States and other countries are moving to a third generation (3G) of mobile telephony.
The defining feature of3G technology is that transmission speeds are significantly faster
than prevailing technology.
A related trend is the growth in use of Wi-Fi (wireless fidelity); these are localized
wireless networks providing high-speed access to the Internet. Whereas 3G could be
described as bringing Internet capabilities to wireless mobile phones, Wi-Fi provides
wireless Internet access for portable computers and handheld devices, such as Personal
Digital Assistants. The two technologies are seen by some as competing for customers
and by others as complementary
providing a broader base and greater choice of
devices for wireless communications and networking.
From the perspective of
spectrum management, a significant difference between the two technologies is that 3G
services operate on designated frequencies licensed by the Federal Communications
Commission (FCC), while Wi-Fi shares unlicenced spectrum with other technologies.
Providers ofthe two technologies share in common the concern that there is insufficient
spectrum available for their services to be developed to full market potential.
—
Industry experts have noted that more efficient uses ofspectrum must be developed
to meet future demand. The U.S. Congress and federal government departments and
agencies are examining the impact that new technology will have on bandwidth demand
and spectrum allocation, prompting Congress to review the policies and laws that guide
spectrum management. Legislation, supported by the Administration, has been
introduced that would make it easier for government agencies to relinquish spectrum to
private wireless carriers for use in providing 3G and other services (H.R. 1320,
Representative Upton and S. 865, Senator McCain).
This report will be updated.
Congressional Research Service + The Library of Congress
CRS-2
Wireless Technology: Development and Demand
In order to deploy third-generation (3G) and other advanced wireless technologies,
telecommunications carriers and their suppliers are seeking effective strategies to move
to new standards, upgrade infrastructure, and develop software for new services. This
migration path includes decisions about using spectrum.
Radio frequency (RF) spectrum is used for all wireless communications. It is
managed by the Federal Communications Commission (FCC) for commercial and other
non-federal uses and by the National Telecommunications and Information
Administration (NTIA) for federal government use. International use is facilitated by
numerous bilateral and multilateral agreements covering many aspects ofusage, including
mobile telephony. Spectrum is segmented into bands of radio frequencies and typically
measured in cycles per second, or hertz.’
Spectrum bandwidth is a finite resource that is infinitely re-usable. Commercial
wireless communications currently rely on bandwidth within a narrow range.
2 American
competitiveness in advanced wireless technology may be constrained by the limited
amount of exploitable bandwidth that is available. This constraint is both specific, in the
inherent finiteness of useful spectrum, and relative, in comparison to the amount of
spectrum available for commercial use in other countries. Developments in technology
have in the past facilitated the more efficient use ofbandwidth within a given portion of
the spectrum. New technologies, such as Software-Defined Radio (SDR) and “smart”
antennae for terrestrial wireless, are being explored and implemented to increase the
efficiency of spectrum and to expand its usable range.
Technology Development. Mobile communications became generally available
to businesses and consumers in the 1980s. This “first generation” technology, still in use,
is analog, the prevailing telecommunications technology of the time. Second generation
(2G) wireless devices are characterized by digitized delivery systems that provide
qualitatively better delivery of voice and small amounts of data, such as caller ID. The
next major advance in mobile technology is referred to as the third generation —3G—
because it represents significant advances over the analog and digital services that
characterize current cellular phone technology. A dramatic increase in communications
speed is the most important technical feature of 3G.
3
Wireless communications services have grown significantly worldwide, and
explosively in some countries. Consumer demand for wireless telephony in the United
States has soared in recent years, totaling over 164 million mobile phone subscribers in
‘One million hertz = 1 megahertz (MHZ); 1 billion hertz = 1 gigahertz (GHz).
2
The FCC limits consideration of bandwidth available for 3G to frequencies below 3 GHz.
The Federal Communications Commission (FCC) identifies key service attributes and
capabilities of3G as the following: capability to support circuit and packet data at high bit rates;
interoperability and roaming; common billing and user profiles; capability to determine and
report geographic position of mobiles; support of multimedia services; and capabilities such as
“bandwidth on demand.” 3G speeds are: 144 kilobits per second at vehicular traffic speeds; 384
kilobits for pedestrian traffic; 2 megabits or higher for indoor traffic, [http:I/wvv.fcc.govI3G].
(Visited May 27, 2004.)
CRS-3
May 2O04. In approximately the same time frame, use of the Internet expanded
dramatically from an arcane tool for specialized research to a popularized, user-friendly
service providing near instant access to information and entertairuuent. Wireless Internet
is widely expected to redefine how computers are used in the future. 3G technologies
bring the wireless Internet revolution to cell phones. Business and consumer demand for
new, advanced wireless services
including 3G and Local Area Networks (LANS),
such as those using Wi-Fi (wireless fidelity)
is considered by many to be an engine
for future growth in American and global economies.
—
—
Third-generation and future developments in wireless technology will be able to
support many services for business and consumer markets, such as: enhanced Internet
links, mobile intranetlextranet, mobile commerce (m-commerce)—including the ability
to make payments—”always on” capabilities, high-quality streaming video and location
identification. In the United States, location-fmder technology for existing cellular phones
is being introduced through the nationwide wireless enhanced 911 (E911) program.
5
Public Policy and 3G
International Agreements on 3G. International agreements that coordinate and
enable global telecommunications are negotiated under the aegis of the International
Telecommunication Union (1TU), a specialized agency of the United Nations. Delegates
to the ITU World Radio Conference in 2000 (WRC-2000) agreed that harmonized
worldwide bands for advanced commercial wireless services were desirable in order to
achieve global roaming and economies ofscale. Resolutions voted by delegates ofWRC
2000 encouraged nations to make available some part of one or more of the three
spectrum bands identified in committee (806-960 MHZ, 1710-1885 MHZ, and 2500-2690
MHZ) for use as harmonized spectrum. In the United States, most of the frequencies
within these bands had been allocated for other uses, requiring reassignment and
relocation in order to provide harmonized spectrum to meet WRC accords.
Harmonized Spectrum. The applications of wireless technology are tied to
spectrum. Infrastructure, such as towers, relay stations, and handsets, must be able to
provide communications along pre-designated frequencies. A benefit of harmonization
is to provide common bands of spectrum dedicated to 3G technology worldwide. This
makes it easier for carriers to cover large geographical areas and for the
telecommunications industry to develop 3G hardware and software. Many industry
observers, however, believe that WRC-2000 did not evaluate the practical considerations
of achieving global roaming capabilities and economies of scale through harmonization.
They argue that countries like China and Brazil are using spectrum to develop 3G
technology in bandwidths not covered by the WRC-2000 resolution, and that global
roaming exists today without the benefit of harmonized spectrum.
Policy Decisions in the United States. Following WRC-2000, President
Clinton directed the Secretary of Commerce to work with the FCC, in coordination with
the NTIA, to prepare studies on allocating bandwidth for harmonized spectrum. In
Statistic updated regularly at [http://www.ctia.org].
CRS Report RS2 1028, Emergency Communications: Wireless Enhanced9l 1 (E911) Issues
Update.
CRS-4
response, the NTIA and the FCC issued reports, respectively, on 17 10-1850 MHZ and
2500-2690 MHZ use.
6
FCC Actions. The report provided by the FCC covered spectrum used primarily
by Fixed Service operators for Multipoint Distribution Service (MDS), Multichannel
Multipoint Distribution Services (MMDS) and Instructional TV Fixed Service (ITFS).
As part of the effort to provide additional spectrum for 3G and other new wireless
technology, the FCC subsequently adopted a First Report and Order and Memorandum
Opinion and Order
7 adding a mobile allocation to the 2500-2690 MHZ range. In March
2003, the FCC announced a Notice of Proposed Rulemaking that will probe alternative
uses for underutilized portions of this spectrum. One of the announced objectives of the
8
proposal is to promote broadband wireless.
In line with the actions of the 3G planning group, the FCC allocated new spectrum
for advanced wireless services, a category that includes but is not limited to 3G.
9 The
spectrum bands are two blocks of 45 MHz each of contiguous spectrum at 1710-1755
MHz and 2110-2155 MHz. The 1700 M}{z band spectrum is used primarily by federal
agencies, including the DOD. Plans are for a rapid relocation of federal agencies other
than the DOD from the 1710-1755 MHz band, with a slower relocation plan for
frequencies used by the DOD. The speed and efficacy of this relocation will be impacted
primarily by the ability to fund the costs of relocation. As part of its plan for relocation
within the 2100 MHz band, the FCC is also looking at other bandwidths that might be
0
freed for advanced wireless servjces.’
NTIA Actions and the Department of Defense. In its report, the NTIA
divided the band into two segments: the 17 10-1755 MHz band, already scheduled to be
made available for commercial use,
’ and the 1755-1850 MHZ band occupied by the
1
Department of Defense (DOD) and 13 other government agencies. In particular, the
report addressed the issue ofreallocating spectrum now used by the DOD. The DOD also
issued a report on the subject with different conclusions than those of the NTIA.
12 The
6
“The Potential for Accommodating Third-generation Mobile Systems in the 1710-1850 MHz
Band,” Final Report, March 2001, U.S. Department of Commerce, NTIA
[http://www.ntia.doc.gov/ntiahome/threegI33001/3g3300 1 .pdf]; and “Spectrum Study of the
2500-2690 MHz Band,” Final Report, March 30, 2001, FCC, [http://www.fcc.gov/3G].
FCC 01-256, September 6, 2001, September 24, 2001.
by the FCC on March 13, 2002, by Notice of Proposed Rulemaking and Memorandum
Opinion and Order (FCCO3-56), FCC News Release, “FCC Initiates Proceeding to Facilitate
Wireless Broadband in the 2500-2690 MHz Bands,” March 13, 2003 [http://www.fcc.govJ.
FCC, Second Report and Order, ET Docket No. 00-258 (2002).
9
‘°FCC, Third Report and Order, ThirdNotice ofProposedRulemaking andSecondMemorandum
Opinion and Order, ET Docket No. 00-258 (2003).
The Omnibus Budget Reconciliation Act of 1993 (47 U.S.C. 927) directed the FCC to
allocate frequencies from the 1710-1755 MHz band on a “mixed-use” basis.
12
“Investigation of the Feasibility of Accommodating the International Mobile
Telecommunications (IMT) 2000 Within the 1755-1850 MHz Band,” 9 February 2000, DOD.
CRS-5
NTIA, DOD and others continue to study and debate Defense’s use of spectrum and
possible plans for migration from the 1755 1850 M}{z band to other spectrum ranges.
-
Spectrum Relocation.’
3 In October 2001 the NTIA announced that a new plan
for the selection of 3G spectrum would be prepared with the FCC, the DOD and other
Executive Branch agencies.’
4 After receiving this and other interagency assessments,
the FCC announced the allocation ofadditional spectrum for advanced wireless services.’
5
In mid-2002, the Department of Commerce circulated draft legislation that proposed the
creation ofa Spectrum Relocation Fund. This would make it possible for federal agencies
to recover relocation costs when they are required to vacate spectrum slated for
commercial auction. Previously, the Strom Thurmond National Defense Authorization
Act of 1999 (P.L. 105-261) authorized agencies to accept compensation payments when
they relocate or modify frequency use in order to accommodate non-federal users. It
authorized the NTIA and FCC to develop procedures for this. The NTIA subsequently
ruled that agencies must submit detailed estimates ofcosts. The FCC suggested that these
estimates be included in the auction process for the relevant spectrum; in effect,
commercial bidders would be covering the costs ofrelocation. The Communications Act
of 1934 would need to be modified to permit the agencies access to auction funds, even
ifpart of the proceeds have been earmarked for their use. A bill (H.R. 1320, Commercial
Spectrum Enhancement Act) introduced in the House on March 18, 2003 called for the
creation of a Spectrum Relocation Fund. It was approved by the House Committee of
Energy and Commerce on April 30 with an amendment to clarify that federal spectrum
could be transferred for non-commercial uses, such as public safety, and for uses where
spectrum is not now auctioned, such as unlicenced spectrum. The bill was passed by the
House on June 11, 2003. The bill was passed, with an amendment, in mark up by the
Senate Committee on Commerce, Science and Transportation on June 26. The addition
of a controversial amendment that could benefit Northpoint Communications reportedly
could harm the bill’s chances for passage by the Senate.’
6 Reportedly, the Acting Director
of the NTIA, Michael Gallagher, has written to Vice President Richard B. Cheney
suggesting that the passage of the Commercial Spectrum Enhancement Act would spur
the development of 3G. He is quoted as saying in the letter that the NTIA “would not
support such an action until identification of alternative frequencies for the affected
federal systems has been completed.”
7
‘
See also CRS Report RS21508 Spectrum Management: Special Funds.
An lntra-Government 3G Planning Group (103 GPG) was created, comprising the NTIA, the
4
‘
FCC, DOD, the Office ofManagement and Budget, the Office ofScience and Technology Policy,
and the Department of State.
‘ FCC News Release, “FCC Allocates Spectrum for Advanced Wireless Service and Proposes
Licensing and Service Rules,” November 7, 2002 [http://www.fcc.gov].
Natio1 Journal’s CongressDaily (AM Edition), “Administration Disapproves of Northpoint
16
Funding Language,” November 11, 2003.
‘‘
“Gallagher Recommends Govt Actions to Forward 3G Deployment,” TELECOM A.M.,
(Today’s News), April 23, 2004.
CRS-6
Wi-Fi
Wireless Local Area Networks (W-LANs) operate on unlicenced spectrum, using
radio frequencies in the free 2.4 GHz and 5.4 GHz spectrum bands. A group of standards
for frequency use in these bands is known as the 802.11 family. The 802.llb standard
is currently the most widely used and is commonly referred to as Wi-Fi, for wireless
fidelity. Wi-Fi provides high-speed Internet access for personal computers and Personal
Digital Assistants (PDAs) and is also widely used by businesses to link computer-based
communications within a small area. The current operating radius for Wi-Fi is 300-350
feet. Links are connected to a high-speed wireline (landline) either at a business location
or through HotSpots. HotSpots are typically located in homes or convenient public
locations, including many airports and café environments such as Starbucks. Another
standard for wireless Internet is Bluetooth, which has a shorter range than Wi-Fi but
works well in cell phones. Bluetooth handles both voice and data; Wi-Fi is mostly data
only, with some use of Voice over Internet Protocols (VoIP). Many industry experts
predict that Wi-Fi and its successor technologies will be the primary link for wireless
Internet, while 3G mobile phones and Bluetooth-enabled PDAs will provide other
th
8 Legislation was introduced in the 108
services.’
Congress (S. 159) that would require
the FCC to allocate “not less than an additional 255 megahertz of contiguous spectrum”
in the 5 GHz band. In response, the FCC has provided 255 MHz of spectrum between
5.47-5.725 GHz, adjacent to one of the frequency bands now used by Wi-Fi.’
9
Conclusion
The continued growth in demand for bandwidth for private and public sector use has
prompted Congress to review the policies and laws that guide the management of this
Some future legislative initiatives regarding spectrum policy might be
resource.
influenced by the formation of a Spectrum Policy Initiative by President George W.
° Among the issues are consideration ofthe law, policies and rulings for spectrum
2
Bush.
allocation that will best meet the sometimes conflicting objectives of protecting
consumers, fostering new technology, encouraging efficiency, bolstering international
competitiveness, and promoting competition, fairness, and access in domestic markets.
18
Financial Times, “Wi-Fi and 3G; cheaper, faster and actually here,” March 12, 2003.
19
FCC, “Revision.. of Rules to Permit Unlicenced.
No. 03-122, released November 18, 2003.
.
20
.
.
Devices in the 5 GHz Band,” Docket
By Executive Memorandum, June 5, 2003 [http://www.whitehouse.gov/news/releases/
2003/06/20030605-5.html]. (Viewed May 27, 2004.)
Government Computer News
The battle for spectrum
030707
By Bob Brewin
Growing demand and a fmite supply of radio frequencies could force
and compromise
confrontation
between the U.S. government and
competing global interests
—
—
Radio frequency spectrum isn’t something you can see, touch or feel, but like other commodities
in short supply, the Earth’s radio frequencies are increasingly valuable as economic and strategic
resources.
The organization that decides how spectrum should be allocated for different purposes is the
International Telecommunication Union (ITU), a United Nations agency. It will host the World
Radiocommunication Conference this fall to allocate global radio frequencies among competing
interests
a highly competitive event that insiders refer to as the Wireless Olympics.
—
Spectrum issues include the demands of an emerging fixed-wireless broadband industry, the
performance of ship and airborne military radar, and the desire of shortwave broadcasters to
operate in a frequency that the U.S. federal government wants to use for data communications
for military and homeland security purposes.
Spectrum is the raw material of mobile and cellular networks that have transformed the way
people communicate. It has liberated the phone from its wired tether and made anywhere,
anytime communication possible for teenagers and millions of people in China, the world’s
largest mobile phone market.
That finite resource has spawned armies of BlackBerry-toting executives and road warrior
students tapping Wi-Fi hot spots in airports and coffee shops to check e-mail and browse the
Web. Remote-sensing satellites use a slice of the spectrum to monitor the weather. Massive
arrays of satellite dishes dedicated to radio astronomy rely on radio spectrum to probe the
mysteries of the cosmos at the beginning of time.
But spectrum remains a limited resource, and authorities must allocate it to meet the
requirements of emerging communications technologies and, at the same time, protect the radio
frequencies that older forms of defense and public-safety communications require.
World radio conference Every three to four years the ITU hosts the World
Radiocommunication Conference for delegations from its 191 member states and 600 sector
members. The sector delegates represent telecommunications companies, such as AT&T, and
digital technology companies, such as Intel. The conference is the venue in which the ITU sets
policies for allocating radio frequency spectrum.
The ITU held its last world conference in 2003. At that time, it established global allocations for
Wi-Fi spectrum in the 2.4 Ghz and 5 GHz bands, a decision that transformed Wi-Fi into a global
phenomenon and enriched numerous companies, including Intel, the Linksys division of Cisco
Systems and Wi-Fi chipset manufacturers Atheros and Broadcom.
The world conference this fall in Geneva is one that Richard Russell, associate director of the
White House Office of Science and Technology Policy, calls the Wireless Olympics because of
the many competing spectrum demands that its participants must resolve. Russell will lead the
U.S. delegation as its ambassador.
The 2007 World Radiocommunication Conference will make spectrum allocations that are
critical for new and emerging technologies of economic importance to the U.S. high-tech
industry. “This is our opportunity to move the ball forward on new services [backed] by U.S.
companies, which will be good for the economy,” Russell said.
In his role as leader of the U.S. delegation, Russell said he must balance U.S. industry interests in
spectrum’s economic value against national security and other public needs. The National
Telecommunications and Information Administration, which defines those public requirements,
has raised questions about new services and technologies that impinge on spectrum that DOD
says it needs for strategic and tactical command-and-control systems. The process The Federal
Communications Commission is responsible for developing initial industry positions for the U.S.
delegation to the conference. For that decision-making, the FCC relies on the World
Radiocommunication Conference 2007 Advisory Committee and five informal working groups
to assess spectrum needs. Those groups held their first meetings in February 2004.
During that preconference period, the working groups approve drafts and present them to the
conference advisory committee. The advisory committee forwards the drafts that it approves to
the FCC, which publishes them for public comment. The consensus reached after the public
comment period will inform the positions that the United States takes at the world conference.
With the conference eight months away, the United States is far from reaching consensus on
many contested issues. Russell said a hot question will be who gets to operate in the 698 MHz to
806 MHz radio frequency band.
That question appeared to be settled after participants at the 1992 and 2000 world conferences
set aside those spectrum bands for broadband wireless mobile or cellular services. Since then,
however, an emerging fixed-wireless broadband industry
with backing from Intel
has
decided it needs to operate in those same frequencies. It wants to offer the public a high-speed
wireless alternative to high-speed Internet service that cable and phone companies now provide.
—
—
The FCC is the present battleground on which Intel and its allies are fighting existing wireless
carriers
AT&T, Verizon, Sprint Nextel and others
to influence the U.S. position on use of
—
—
that high-speed radio frequency band. Nancy Victory, a partner at the Washington, D.C., law
firm Wiley Rein and chairwoman of the FCC advisory committee, said the two industry sides
appear to be so far apart that the U.S. delegation might go to the world conference with two
opposing views on the use of the 698 MHz to 806 MHz band. Russell, however, said he is
confident that the United States will end up with a unified position on the contested spectrum
that will satisf’ all industry segments before he heads to Geneva.
While that question is debated in the months preceding the conference, DOD is concerned about
how the ITU will rule on frequencies that the department uses for navigation and mobile
communications. The ITU must decide whether to let industry use those spectrum bands for new
broadband mobile services. Badri Younes, DOD’s director of spectrum management, said
allowing broadband mobile services to operate in those frequencies could degrade the
performance of ship and airborne radars and mobile communications systems.
The frequencies in which DOD has a stake are the 410 MHz to 430 MHz band that fixed and
older mobile communications systems use, the 2700 MHz to 2900 MHz band on which pilots
rely for radio navigation, and the 3400 MHz to 3650 MHz radio frequency band that ship and
airborne radars use.
DOD also has an interest in whether the ITU decides to protect high-frequency (HF) spectrum
from encroachment by shortwave broadcasters worldwide. All three military services rely on HF
for long-range terrestrial communications. Fighting to retain control of the HF bands might seem
like a last-ditch stand to keep horse-drawn buggies, some spectrum experts say. Broadcasters,
such as the BBC, have moved from HF and now rely on satellite and Internet feeds.
But the development of high-quality digital radio broadcasting technology has changed the
picture and made HF suddenly more appealing. Shortwave broadcasters worldwide want the ITU
to allocate to them as much as 800 KHz of additional spectrum in the HF band from 4 MHz to 10
MHz. A new golden age Analog shortwave broadcasts are typically interrupted by hisses and
pops. But the new technology, Digital Radio Mondiale (DRM), will enable shortwave
broadcasters to transmit programs thousands of miles with the signal quality and audio clarity of
FM radio, said Nigel Holmes, transmission manager at Radio Australia.
DRM could lead to a golden age for shortwave broadcasting, Holmes said.
Broadcasters worldwide
the BBC, Deutsche Welle in Germany, Radio France, Radio Sweden,
Radio Canada and Radio New Zealand
have embraced it. The National Association of
Shortwave Broadcasters in the United States, most of whose members are religious broadcasters,
supports DRM, and it has urged the FCC to push for allocating additional HF spectrum for
broadcasting.
—
—
But Eric Johnson, professor of engineering at New Mexico State University and a member of the
AFCEA International HF Industry Association, said HF is equally well-positioned to support a
golden age of data communications. Thanks to new standards, some of which Johnson helped
develop, HF can support data rates of 9,600 bits/sec in a 3 KI-{z channel and 64 bits/sec in a 12
KHz channel.
In an era in which users can access the Internet at speeds measured in megabytes of data, 64
bits/sec does seem like horse-drawn buggy speed, but it exceeds the data rate of satellite
communication systems on some of the Navy’s smallest ships, Johnson said. The Air Force
already relies on HF in its High Frequency Global Communications System to transmit e-mail
messages to airlifters the Air Mobility Command operates.
The Homeland Security Department’s Customs and Border Protection organization operates a
nationwide HF network to communicate with its boats, planes, helicopters and vehicular assets,
Johnson said. Australia and the United Kingdom recently installed new HF networks. Those
DHS and overseas networks could be undermined if shortwave broadcasters receive expanded
HF spectrum allocations, Johnson said.
Younes said HF spectrum is critical to DOD’s combat operations in Afghanistan and Iraq and
relief operations such as those DOD conducted in late 2004 and early 2005 after a tsunami hit
many Asian and Indian Ocean countries. DOD must protect HF spectrum from broadcasters’
incursions as it develops new Internet-based applications for HF, he said. HF is essential for
communications with partner countries in multinational coalitions, many of whom use it as their
sole means of communications, he added.
The FCC has deliberated but has not reached a decision about whether to back the shortwave
broadcasters or the government’s tactical and mobile users of HF spectrum. Johnson said the
FCC needs to side with DOD and other federal stakeholders and do so well ahead of the
conference to protect HF spectrum that DOD and other federal agencies use. Many countries are
waiting to learn the U.S. policy position before they solidify their own, and an early signal from
the United States could slow the broadcasters’ momentum in seeking additional HF spectrum, he
said.
Remote-sensing applications Another contested frequency is the 36 0Hz to 37 GHz band. The
National Academy of Sciences, for example, wants the United States to push for mandatory
power limits on communications satellites operating in that band to protect satellite-sensing
systems in the Earth Exploration Satellite Service. U.S. government and nongovernmental
entities operate that service.
The academy informed the FCC that scientists rely on that band for passive sensing of the
Earth’s surface and atmospheric conditions. Scientists say they have little flexibility in choosing
frequencies to conduct their remote-sensing observations because the specific frequencies of
observed elements, such as the absorption and emission of passive microwave radiation, are
established by the laws of physics and chemistry.
But the Satellite Industry Association, which represents satellite operators, service providers and
manufacturers, has told the FCC that it opposes mandatory power limits on communications
satellites. Any such limits would harm the development of new services and technologies that the
satellite industry wants to offer, it said.
The competing demands don’t end there. The airline industry and the Federal Aviation
Administration have put in their bid for additional spectrum for flight test telemetry. Boeing says
it needs access to additional radio frequencies to support a near-doubling of the test data that the
FAA now requires before certifying commercial aircraft.
Boeing says it monitored 64,000 test points in 1995 to certify its 777 airliner. It expects to
monitor 100,000 test points for its new fuel-efficient 787 Dreamliner series aircraft that will
carry 250 to 330 passengers. Younes agreed that a failure to gain additional spectrum for flight
testing would inevitably cause schedule delays and cost increases for Boeing and other airplane
manufacturers.
Russell said his goal in representing the U.S. is to be competitive in that and other high-profile
spectrum contests at the Wireless Olympics, where going for the gold will mean more to the
United States than winning a coveted medal.
© 1996-2008 1105 Media, Inc. All Rights Reserved.
LO5to,, ni
he oston 1obc
Big wireless carriers get set to free the phone
By Carolyn Y. Johnson, Globe Staff
I March 28, 2008
It is a scenario rarely seen in today’s technology market: Cellphone customers wander into any store, pick
any device from a shelf, and connect it to any wireless network one as open as the Internet.
-
But spurred by growing demand and a federal airwaves auction that closed last week, the major wireless
carriers are stepping away from a model in which each cellphone is controlled by a single company that
sells customers a device locked to their network, demands a lengthy contract, and limits the phone’s
features.
‘We call it a 180-degree about-face to open access, and it really is a fundamental shift in business
models if we can get there,” said Linda Barrabee, senior analyst at Yankee Group, a technol ogy
research firm, and author of a recent report, “Open Access is the New Black.”
-
Just a year ago, open celiphones were largely the province of hackers or frequent international travelers.
But the newest website educating people about how to connect outside devices to a network is no hacker
forum: It is run by AT&T, the nation’s largest wireless carrier. Verizon Wireless, the number two provider,
held a two-day conference this month to tell independent developers how they could get their devices
certified to work on its network. And Google, the Internet giant that lobbied for open access and last year
unveiled an open mobile platform called Android, this week detailed a proposal to use portions of the
airwaves between television channels for open Internet access.
Suddenly, open access is fashionable.
The pivotal moment occurred in the run-up to this year’s federal auction of valuable broadcast spectrum,
which can also be used to carry telephone calls and data. Because of federal requirements that TV
channels broadcast digital signals by February 2009, some airwaves now being used for TV will be
available for other uses. As the auction of what is often called the last beachfront property in the wireless
world approached, politicians, consumer advocates, and Google launched a discussion about the future
of telecommunications in the United States.
The Federal Communications Commission attached rules to a slice of the spectrum, requiring that the
winner of that portion leave the network open to outside devices and applications.
Meanwhile, more capable phones and networks meant people began to expect the devices to function
more like computers, and public awareness of restrictive carrier policies increased.
When the high-profile iPhone was launched exclusively on the AT&T network last summer, it increased
public awareness that devices were locked, meaning they were only compatible with one carrier. People
who paid full price for an iPhone could not use it through another carrier, even if the carrier T-Mobile, for
instance used compatible technology.
-
-
When the iPhone was unlocked last summer to allow use with other carriers, what would have seemed to
be an insider-geeky event drew widespread attention and national news stories.
Last fall, SDrint agreed to provide codes to unlock its phones as part of a class-action settlement. And all
four major carriers have said that when a customer ends a contract early, termination fees will be
prorated.
Driven by a combination of regulation and public awareness, the zeitgeist for “open” was born. Earlier this
month, Apple began allowing outside software developers to create their own programs, such as games.
Not long ago, carriers bristled at the notion that the nation’s wireless system stifled innovation. Verizon
Wireless fought the FCC’s decision to attach open access requirements to the spectrum.
“Imposing any such requirements in the competitive wireless market would reduce the revenue the
government will receive from the spectrum auction and limit the introduction of new and innovative
wireless services,” the company said in a statement in July, as the rules were being debated.
But Verizon Wireless eventually accepted the rules, and earlier this month it won the coveted block of
spectrum subject to the open-access conditions. Also, AT&T and other companies won other portions of
the spectrum, and the auction raised $19.1 billion. At its conference for developers last week in New
York, Verizon emphasized that new partnerships were essential for innovation.
‘We believe that as the two great megatrends of mobility and the Internet come together, the next wave of
growth will come from a whole new generation of devices, applications, and services,” said Ivan
Seidenberg, chief executive of Verizon Communications Inc., which owns 55 percent of Verizon Wireless.
“No single company whether you’re a carrier, a manufacturer, a software company, or anybody else
will be able to envision all these uses or meet all the needs on their own.”
-
-
Not to be outdone, AT&T launched a website the day before the conference, emphasizing its open
network.
“This is not new business for us,” said Mark Collins, vice president of Consumer Data Products for AT&T.
The company uses GSM, the predominant wireless technology and “the most open standard on the
planet,” he said.
Still, Collins said AT&T phones would be sold locked -and unlocked only upon customer request. T
Mobile has a similar policy. Sprint and Verizon Wireless use different technology. Sprint and T-Mobile
also moved to embrace openness as members of the “Open Handset Alliance.” To some extent, the
technology for open systems has been available all along, but wireless carriers have created a different
model in the United States than in Europe or Asia, where consumers can easily swap a phone between
providers.
But by the end of this year, Verizon will allow approved devices that aren’t offered by the company to
connect to its network. People can already buy phones and add T-Mobile or AT&T service. Also, Sprint is
beginning to roll out its Xohm mobile broadband WiMAX service, touted as part of its “open Internet
vision.”
Ultimately, loosening the relationship between the wireless provider and devices could spur innovation
devices that phone companies may never have developed. Sprint, for instance, powers the wireless
connection for the Amazon Kindle, an e-book reader.
-
Openness comes at a pivotal point for the industry. Companies are struggling to prevent what happened
online, when people moved away from proprietary “walled gardens” like AOL and Prodigy and onto the
wider Web.
As the cellphone market becomes saturated, providers will have to find ways to grow other than selling
people their first phone, and as data services become more common, “their whole business model is
called into question,” said Paul DeBeasi, senior analyst with the Burton Group. “If instead of paying for
minutes people pay for Skype that’s what happened to the old AT&T. People stopped paying for long
distance; it became a commodity.”
-
But even as companies try various approaches to openness, their definition of “open” remains unclear.
Are they paying lip service to a popular idea or will they open a new frontier in phones and computing?
-
According to Barrabee, it will be years before the mobile ecosystem really changes and technological
hurdles are overcome so people can really walk into stores and buy any device and pick any network.
And consumers may continue to prefer the current US model, where it is clear whom to call for customer
support and people are used to cheap phones and long contracts.
“The concept of open access is a bit intoxicating, but we really need to see the details and action need
to see an ecosystem emerge and then see distribution channels for these devices and services, before
we see the impact,” said Barrabee. “And then it is really a question how much does anybody want?”
-
-
Carol n Y. Johnson can be reached at c ohnson
© CoDvriaht 2008 The New York Times Company
lobe, corn..
RCR fie1ess News
WiMAX to snag 47M subscribers by 2013
Story posted: May 9, 2008 8:18 pm EDT
-
A new report from Juniper Research predicts WiMAX growth will begin to take off in the period between 2009 and
2011, growing to more than 47 million subscribers worldwide by 2013.
The report predicts W1MAX will find a strong foothold as a substitute for DSL technology.
“We determined that the vast majority of the W1MAX 802.16e trials and network contracts which are being announced
almost daily will begin by providing fixed broadband,” said Howard Wilcox, author or the report.
“WiMAX can deliver broadband not only to uriwired areas, but can also improve speeds for subscribers who are on
the fringe of DSL coverage in metropolitan areas.
“We anticipate that mobile usage will develop after initial demand for fixed and portable services
is a flexible platform that can operate in all three modes of usage,” said Wilcox.
—
WiMAX 802.16e
The report predicts global revenues for WiMAX service as a DSL replacement technology will exceed $20 billion a
year by 2013 arid that 12% of the DSL subscriber base at that time Will be replaced by WiMAX technology. The Far
East, North America, Western Europe and Africa/Middle East will be the largest markets for WIMAX as a DSL
substitute, predicts the report.
The report cautions that uptake could be slowed by a lack of suitable devices and timely network construction.
PRINTED FROM: http:/Iwww.rcmews.comlappslpbcs.dIIIa,ticle?AID=/20080509/SUB/94007759211008/newsletterl 6&template=pnntart
Entire contents © 2008 Cram Communications, Inc.
Google to bid for U.S. mobile airwaves
Fri Nov 30, 2007 4:22pm EST
By Eric Auchard and Peter Kaplan
SAN FRANCISCO/WASHINGTON (Reuters) Google Inc said on Friday that the company would bid on
coveted airwaves to launch a U.S. wireless network, pitting it against established telecommunications
players AT&T and Verizon.
-
The Internet leader said in a statement that it was ready to go it alone rather than rely on partners in
bidding in the Federal Communications Commission-run auction of 700-megahertz wireless spectrum
due to begin on January 24.
The Silicon Valley-based company said it would make its filing ahead of the FCC deadline on Monday
for companies to declare their interest in joining the airwaves bidding. “We believe it’s important to
put our money where our principles are,” Chief Executive Eric Schmidt said. “Consumers deserve more
competition and innovation than they have in today’s wireless world.”
Wall Street investors have reacted cautiously to Google’s latest move to expand beyond its core Web
search and online advertising franchises, worried the potential upfront costs and eventual network
build-out could exceed $10 billion.
But some analysts have speculated that Google was more interested in ensuring certain requirements
for network openness and that it was bidding just to preserve those rules.
“The real question here is whether Google’s intent is to bid up to the reserve price and assure that the
openness condition stays in place,” Stifel Nicolaus analyst Blair Levin wrote to investors. “Or is the real
purpose to actually win?”
And despite the excitement surrounding a Google bid, Stifel Nicolaus said in a research note that it
suspects Verizon will probably end up winning the auction for the C block spectrum. Google shares
closed down $4.00 at $693.00 on Nasdaq.
Bidding separately instead of assembling a coalition does not rule out Google later signing up partners
if it wins the bidding, said a source familiar with the company’s strategy. But the FCC has “anticollusion” rules that prevent deal-making between potential bidders during the auction period.
The source said Google was eyeing the biggest chunk of spectrum up for auction
the “C block”
considering
bidding
but also was
on separate spectrum reserved for public safety agencies but which
will allow some commercial uses. A Google spokesman declined to comment on the company’s
bidding strategy.
--
--
LAST CHANCE FOR NEW PLAYER
The auction is expected to take several weeks, or even months, of daily, back-and-forth bidding, with
the identities of the bidders kept secret. Big spectrum bidders typically draw up elaborate strategies,
often with input from game-theory experts.
Expected bidders include AT&T Inc and Verizon Wireless, the No. 1 and No 2. U.S. wireless network
operators. Verizon Wireless is a joint venture of Verizon Communications Inc and Vodafone Group Plc.
Less certain are the strategies of satellite broadcasters DirectTV and EchoStar Communications and
cable networks Comcast Corp and Time Warner Inc as well as other wireless players Sprint, T-Mobile
and Clearwire, according to Leviri.
AT&T is in merger talks with EchoStar that could lead them to join forces in the bidding, but with the
deadline looming the time to strike such a deal is short, the Stifel Nicolaus analyst said.
These radio waves are being returned by broadcasters as they move from analog to digital signals
early in 2009. The signals can go long distances and penetrate thick walls. The auction is seen as a
last chance for a new wireless player.
Goole and other Silicon Valley leaders see the wireless spectrum as a way to create more open
cornpetition for mobile services and devices than existing networks
putting the industry on a footing
similar to the free-wheeling Internet.
--
The company won some changes in rules governing use of the spectrum several months ago, but was
denied other requests, including a rule that would have required winning bidders to resell access to
their spectrum on an open wholesale basis.
The winning bidder must provide open access to any device consumers choose to use on the network
if the reserve price of $4.6 billion for the “C Block” is met at auction, Google said. If the reserve price
is not met, the auction would be rerun without the so-called “open-platform” conditions.
On Tuesday, Verizon Wireless announced it had acceded to Google’s open-platform demands and
would open its network to any phone or software application by the end of 2008. But Levin said
Google may still want to bid high enough to lock in a government-enforced open-platform condition on
the 700-megahertz spectrum.
If its bid proves successful, Google could operate a wireless network itself or seek partners to help it
build out the network and to potentially resell wireless services.
Google’s announcement was greeted as good news at the FCC, said a source at the agency. “It means
that they’re willing to go a little bit further into the water,” the source said. “They’re not just dipping
their toe anymore.” FCC officials hope the company’s participation will mean a possible new player in
the wireless business and boost the amount of money the government can bring in from the auction.
“We know we’re going to have somebody in the mix who has a lot of money and who is showing signs
in being interested in winning,” said the source.
Supporting Reference for Monetization
To Fund Investment Opportunities
Many cities and municipal agencies purchase government bonds due to their low risk. As the information
provided in this package has shown, cell site leases or cell sites options are the opposite, high risk. This is
due to the technological and economic factors that are affecting the telecommunications industry today.
Below we have shown an example of the situation that we are advising our clients to take, cash in on the
certain net present value of your site lease options and purchase government bonds. This situation will
eliminate your risk of the termination and bring in slightly more income.
ATS Communications
Initial Rent
EscalatIon
i-lnanciai Monetization Moael
‘r
170000
1,75100
1,803 53
1,85764
1,91336
1,970.77
2,029.89
2,090.79
2,153.51
2,218.11
2,284.66
2,353.20
2,423.79
4
7
T
T
•“T
T
W
...T
‘T
22
25
W
T
r
ö
$
S
S
54
....2j.00
16
17
“‘
,
3%
A,nuaIRevenue1
1
2
19
70000
L
1
$
.
S
$
S
$
$
S
$
$
$
$
$
2,894 1
2,980.96
3,070,39
3,162.50
3,257.38
3,355.1
3,455.7
3,559,4
3,868.21
3,776.19
3,88948
4,006.16
-
20,40000
61200
21,01200
63036
21642 36
649 27
22,291 63
66875
22,96038
68881
23,649.19
70949
24,358.67
76
75268
25,08943
25,84211
fl526
26,61737
798o2
27,41589
48
28,23837
‘iV iS
29,08552
672o7
29,95809
89874
30,85683
“70
31,78254
32,73601
628
33,718 09 $ ‘i
34,72963 $ 1.041 89
35,771 52 51,..... .5
36,84467 S 1 .0534
37,95001 51.13850
39,08851 VT..... ..6
40,261 16 F”T,207 83
41,46900 $124407
42,713 07 S 1,28139
43,99446 $ 1,31983
45,31430 S 1.35943
46,67372 S 1,40021
48,073.94 $ 1,44222
$
S
S
S
S
21,01200
21,64236
22,291.63
22,960.38
23,64919
24,358.67
25,089.43
25,84211
26,61737
27,41589
28,23837
29,08552
29.95809
30,85683
31,78254
32,73601
33,71809
34,729 63
35,771_52
36,844 67
37,95001
39,08851
40,261 16
41,46900
42,71307
43,99446
45,31430
46,67372
48,07394
49,51615
Year
—
—
—
—
—
18700000
197,75250
209,123 27
221,14786
233,86386
247,31103
261,531 41
276,569.47
292.47222
289.37
—
3,fl73 51
4
.
111,55556
—
—
—
—
—
.
$
240,87903
,u
‘‘T
12
13
—
—
T
$
390,800.38
.
-
‘
$
564,600 31
.
$
766,08206
‘(5
...2.
17
‘i’W
19
20
21
22
23
24
25
.
,
T
—
—
—
—
386,800.03
....3
69
‘io,,i4
‘4
—.,...-.
‘...
511,55034
D40,Ub4 49
Q1.UQ
4
—
—
9
-.98
—
...
—
715,43574
—
80..,..7 26
—
..0 28
5: 946...7..7
—
$
999,654.63
]
.....,
0.0575
0.0575
0.0575
0.0575
00575
0.0575
0.0575
0.0575
0.0575
0.0575
0.0575
0.0575
0.0575
0.0575
0.0575
0.0575
0.0575
0.0575
00575
0.0575
0.0575
0.0575
0.0575
0.0575
0.0575
0.0575
0.0575
0.0575
0.0575
0.0575
WY.ar RevflUaI $
HV.
.
-.“-
“
‘.-.
$187OOOO0
5-75%
‘
AnnuaI mom. Ending Balance
8
.
S
Purchas.’Prlc.
3OYr,Treasury
B9965463
!
S’ .187000.O0(
264 Lamb.4 St., Suffe 402, LIke Forest, CA 92630, P551.: (949) 305-7648 FIX (949) 766-6984
Webkte www ebconwn.oom BiieO: InfoCetIces cn
10.75250
11,37077
12,02459
12,71600
13,44717
14,22038
15,03806
15,90274
16,81715
17,78414
18,80673
19,86811
21,031 68
22,241 00
23,51986
24,872.25
26.30241
27,81479
29,41414
31,10548
32,89402
34,78543
36,78559
38,90076
41,13756
43,502 96
46,00438
48,64964
51,44699
54,405.19
S
$
S
S
S
$
—
197,75250
20912327
221.147 86
233,86386
247,31103
261,53141
276,56947
292,47222
309,28937
327,07351
345,88023
365,76835
386,80003
409,041 03
432,56089
457,43314
483,73554
511,550 34
540,96448
572,06994
604,96396
639,74939
676,53498
715,43574
756,57329
800,076 26
846,08064
894,73028
946,17727
1,000,58247
To Fund Capital Improvements
Receiving a lump sum of the net present value of the monthly rent/income eliminates the risk of lease
termination, which can take place within 30 days. In actuality, the Communication Site Lease Agreement
that you are holding is nothing more than a 30 day option, granting the “Lessee the authority to determine
if the Premise is not appropriate for its operations for economic, environmental or technological reasons.”
Many times the revenue generated from this lease agreement or option is used to fund necessary capital
improvements within parks, city streets, and city buildings. But with the 30 day termination clause
embedded in the lease or option there is a high probability that the payout of 30 years will never come due
to the economic and technological changes previously explained. Along with the risk of waiting for this
stream of income to come through on a monthly basis, there is inflation and the rising cost of materials to
account for, which exponentially increases the cost of capital improvement projects as time goes on.
I
I
THE ASSOCIATED GENERAL CONTRACTORS OF AMERICA
2300 Wilson Blvd., Suite 400 •Arllngton, VA 22201
Title: RISING COSTS, UNCERTAIN DEMAND SQUEEZE NONRESIDENTIAL CONSTRUCTION, AGC SAYS
Date: March 18, 2008
Washington, D.C. — “Construction costs continued to outpace other inflation measures in February, while
demand softened for some project types,” Ken Simonson, Chief Economist for The Associated General
Contractors of America (AGC), said today. Simorison was commenting on two new economic releases for
February: producer price indexes (PPIs) from the Bureau of Labor Statistics (BLS) and housing permits from
the Census Bureau.
“The PPI for inputs to construction industries—materials used in all types of construction plus items
consumed by contractors, such as diesel fuel—climbed 0.6 percent in February, compared to 0.2 percent for
the PPI for finished goods and 0.3 percent for the consumer price index (CPI), before seasonal adjustment,”
Simonson observed. “That continues a trend since steel prices first jumped at the end of 2003. From
December 2003 through February 2008, prices for these construction inputs have soared 31 percent, vs. 15
percent for the CPI.
“That huge gao is especially troublesome for contractors on public projects.” Simonson asserted. “Public
agencies often rely on the CPI to proiect future costs but they are coming up short of the dollars needed to
award contracts. The roblem is most acute with highway orolects. where the huae run—ups in diesel.
asphalt, concrete and steel costs have pushed up the PPI by 50 percent since December 2003
“In February, there were outsized increases in the PPIs for copper and brass mill shapes (5.8 percent); hot—
rolled bars, plates and structural shapes for rebar and structural steel (3.5 percent) and diesel fuel (2.2
percent),” Simonson noted. “Still worse, all of these materials have risen even more since the PPI data was
collected in mid—February. Yesterday, the Energy Information Administration reported that on—highway
diesel costs $3.97 a gallon nationally, up 69 cents in the five weeks since the PPI reporting date and up
$1.29 or 48 percent from a year ago.
“Meanwhile, demand is falling for multi—unit residential projects,” Simonson added. “Census reported today
that multi—unit permits plunged 11% from January and 23% from February 2007. Demand for office, hotel
and retail construction is reportedly weakening as well.
“Nevertheless, I do expect continued strength for hospital, university, power, energy and communication
construction,” Simonson concluded. “There is ongoing demand for these facilities, and their financing is
generally more secure than for projects that depend on short—term rents.”
NOTE: A chart comparing the increase in the PPI for construction inputs and the CPI is at
www.agc.org/febppi. For a more detailed discussion of trends in construction activity, materials costs, and
labor, see AGC’s Construction Inflation Alert at www.agc.org/cia.
The Associated General Contractors of America (AGC) is the largest and oldest national construction trade
association in the United States. AGC represents 33,000 firms, including 7,500 of America’s leading general
contractors and 12,500 specialty—contracting firms. More than 13,000 service providers and suppliers are
associated with AGC through a nationwide network of chapters. Visit the AGC Web site at www.agc.org. AGC
members are “Building Your Quality of Life.”
be uork imc5
nyt mes
corn http://www.riytimes.com/
January 26,2008
Building Costs Deal Blow to Local Budgets
YARDLEY
By
SEATTLE
—
State and local governments in many parts of the country are struggling to pay for roads, bridges
and other building projects because of rising construction costs, adding another burden to budgets already
stressed by the troubled housing market.
The problems have come as many governments pursue ambitious projects to improve roads and airports,
build schools and upgrade long-neglected water and sewer systems. Many of the projects were conceived
when money from property, sales and income taxes was steady and interest rates low, but officials say the
ground has shifted beneath their feet.
“Everybody’s scared,” said Uche Udemezue, director of engineering and transportation for San Leandro,
Calif., which will soon put out a request for construction bids on a retiree center and a parking garage. “You
don’t know what you’re going to find when you go out to bid.”
Costs have jumped for projects as varied as levee construction in New Orleans, Everglades restoration in
Florida and huge sewer system upgrades in Atlanta. The reconstruction of the Interstate 35W bridge in
Minneapolis, a $234 million project, has been fast-tracked for completion by December, and state officials
say it is too soon to know whether it will come in on budget.
The impact has been felt in different regions at different times, and not every project has been high-profile.
In Oregon, high costs have forced the State Department of Transportation to slow the rate at which it
upgrades roads and bridges. In Seattle, school building projects were put on a fast track this fall because of
fears of cost overruns.
“We escalated our project schedule to get ahead,” said Fred Stephens, director of facilities and construction
for Seattle Public Schools.
Nationwide, increasing costs first became a problem for some projects more than two years ago, and in some
regions the rate of increase has dropped in the past year. But some regions are tighter than ever, and the
pressure from the high costs can be more acute in the context of general revenue declines.
The list of culprits for the increases often depends on the rate of growth and construction in a particular
region, with labor costs playing a role along with the rising prices of materials like steel and concrete, and
asphalt, fuel and other petroleum-based products.
Experts say high costs are linked to competition from a global development boom, particularly in China and
India; the housing boom in the United States; and the rush to rebuild after Hurricane Katrina in 2005 and
other recent hurricanes that struck Florida and the Southeast. In the Northwest, public projects have
competed with downtown construction surges in Seattle and Portland. Just across the Canadian border,
hotels and highways are being built to prepare for the 2010 Winter Olympics in Vancouver.
The costs have added to what has become an increasingly bleak economic forecast for many states and local
governments. At least 25 states expect to have budget deficits in 2009, according to the Center on Budget
and Policy Priorities, which estimates the combined budget shortfall for 17 of the states at $31 billion or
more. Many cities, too, see difficult times ahead as revenues wane and costs increase for wages, pensions
and health care.
“We’re talking about all levels of government being in some revenue constraints at a time when the service
costs aren’t going down,” said Chris Hoene, the director of policy and research for the National League of
Cities.
In some places, the news is not all bad. Recent declines in residential construction are beginning to force
contractors to be more competitive when they bid for government work. Yet some government officials see
that as a dubious silver lining.
In Oregon, low bids for recent bridge projects came in at $18 million, about 10 percent below what the state
had projected. That was unimaginable a year ago, but the relief is relative, said Tom Lauer, the major projects
manager for the Transportation Department.
“We’ve been getting hit so hard that we’ve been pumping them up the last couple of years,” Mr. Lauer said
of the state’s internal cost projections.
“I didn’t get a price break,” he said of the recent bid. “I may just have more predictable pricing. I still can’t
afford to do other stuff.”
In Newcastle, a growing Seattle suburb, the situation is emblematic of the struggles confronting towns and
school districts across the country. Two main goals prompted the improvements now under way on a main
thoroughfare, Coal Creek Parkway. Widening a bottleneck on the road would help relieve congestion on
nearby Interstate 405. And doing it with style
installing landscaped medians between lanes
—
—
using steel on a bridge to evoke an old train trestle and
would send the signal that Newcastle is ready to do business.
Then the bids came back. “Slack-jawed,” said John Starbard, the city manager, when asked his reaction to the
bids.
Mr. Starbard said even the project’s engineering consultant, CH2M Hill, was stunned when what they
believed was a very conservative $38 million estimate in March 2007 was met with a low bid of more than
$44 million for a mile’s worth of road and bridge improvements.
But waiting to build was not an option. The city had already received help from Senator Patty Murray,
Democrat of Washington, and state lawmakers, as well as the State Transportation Improvement Board. It
went back to the board and received $2 million more.
“It was a shared sticker shock, but they had seen this with other projects so they were not as surprised,” Mr.
Starbard said of the board.
In Newton, Mass., a Boston suburb with a population of more than 80,000, the estimate for the new Newton
North High School was $104 million in 2004. Four years later, the foundation is about to be poured and the
estimate is now at least $186 million, said Jeremy Solomon, a city spokesman. Mr. Solomon said about $25
million of the increase involved changes to the original plan, for asbestos abatement, adjustments to the
heating and air-conditioning system and other factors. Otherwise, he said, the increase resulted from rising
building costs.
“We kind of got caught in a period where construction costs grew rapidly,” said Mr. Solomon, citing steel and
fuel costs, among others.
The need for public improvements only grows greater. Costs are rising even as engineers across the country
say infrastructure is rapidly decaying.
In San Leandro, a city of 78,000 in the San Francisco Bay Area, Mr. Udemezue said the city could not afford to
delay work on the parking garage and retiree center.
“We can’t wait,” he said, “because we don’t know if the prices are going to come down or go up.”
In the grading guide known as the Pavement Condition Index, zero is not far from a dirt strip and 100 is a
fresh new roadway. When Mr. Udemezue began working for San Leandro 16 years ago, the average road
ranking in the city was nearly 70. Now it is closer to 60, despite what Mr. Udemezue said were the city’s
efforts to keep up maintenance.
Years ago, there was more money in the city’s general revenue stream that could be diverted to help with
basic maintenance, which Mr. Udemezue said required about $5 million a year.
That general revenue now goes to other needs, like public safety, and the roads go wanting, with flat revenue
from gas taxes and other declines leaving about $1.2 million to maintain roads each year. The $13 million
retiree center and the $8 million parking garage have been affected, too, with the city dropping plans to build
commercial space beneath the garage and reducing the space for social programs in the center.
Mr. Udemezue and others say they have heard that things may be stabilizing, but they cannot be sure.
Even in places where the rise of costs has slowed, said Ken Simonson, chief economist with the Associated
General Contractors of America, “it’s dormant at best.”
Cost overru
From Wikipedia, the free encyclopedia
• Learn more about using Wikipedia for research•
Cost overrun is defined as excess of actual cost over budget. Cost overrun is also sometimes
called “cost escalation,” “cost increase,” or “budget overrun.” However, cost escalation and
increases do not necessarily result in cost overruns if cost escalation is included in the budget.
Cost overrun is common in infrastructure, building, and technology projects. One of the most
comprehensive studies [1] of cost overrun that exists found that 9 out of 10 projects had
overrun, overruns of 50 to 100 percent were common, overrun was found in each of 20 nations
and five continents covered by the study, and overrun had been constant for the 70 years for
which data were available. For IT projects, an industry study by the Standish Group (2004)
found that average cost overrun was 43 percent, 71 percent of projects were over budget, over
time, and under scope, and total waste was estimated at US$55 billion per year in the US alone.
Spectacular examples of cost overrun are the Sydney Opera House with 1,400 percent, and the
Concorde supersonic aeroplane with 1,100 percent. The cost overrun of Boston’s Big Dig was
275 percent, or US$1 I billion. The cost overrun for the Channel tunnel between the UK and
France was 80 percent for construction costs and 140 percent for financing costs.
Three types of explanation of cost overrun exist: technical, psychological, and politicaleconomic. Technical explanations account for cost overrun in terms of imperfect forecasting
techniques, inadequate data, etc. Psychological explanations account for overrun in terms of
optimism bias with forecasters. Finally, political-economic explanations see overrun as the result
of strategic misrepresentation of scope and/or budgets.
All of the explanations above can be considered a form of risk. A project’s budgeted costs
should always include cost contingency funds to cover risks (other than scope changes imposed
on the project). As has been shown in cost engineering research [1], poor risk analysis and
contingency estimating practices account for many project cost overruns. Numerous studies
have found that the greatest cause of cost growth was poorly defined scope at the time that the
budget was established. The cost growth (overrun of budget before cost contingency is added)
can be predicted by rating the extent of scope definition, even on complex projects with new
technology.
Cost overrun is typically calculated in one of two ways. Either as a percentage, namely actual
cost minus budgeted cost, in percent of budgeted cost. Or as a ratio, viz, actual cost divided by
budgeted cost. For example, if the budget for building a new bridge was $100 million and the
actual cost was $150 million then the cost overrun may be expressed as 50 percent or by the
ratio 1.5.
COMMODITY PRICE DATA
Commodity
Energy
Coal, Australia
Crudeoil,avg,spot
Crude oil, Brent
Crude oil, Dubai
Crudeoil,WestTexaslnt.
Natural gas Index
Natural gas, Europe
Natural gas, US
Natural gas LNG, Japan
Unit
Annual averages
Jan-Dec Jan-Dec Jan-Apr
2006
2007
2008
Quarterly averages
Jan-Mar Apr-Jun Jul-Sep Oct-Dec Jan.Mar
2007
2007
2007
2007
2008
Monthly averages
Feb
Mar
Apr
2008
2008
2008
j S/mmbtu
$/mmbtu
S/mmbtu
49.09
64.29
65.39
61.43
66.04
181.6
8.47
6.72
7.08
65.73
71.12
72.70
68.37
72.28
186.5
8.56
6.98
7.68
116.25
98.67
100.05
94.35
101.61
243.4
11.19
9.02
10.55
53.19
57.23
58.07
55.58
58.03
187.9
8.51
7.23
6.95
57.91
66.13
68.73
64.71
64.96
186.1
8.00
7.50
7.14
68.37
73.50
75.04
69.97
75.48
174.4
8.34
6.17
7.68
83.47
87.61
88.95
83.21
90.67
197.7
9.37
7.03
8.96
114.00
95.31
96.67
91.30
97.94
235.0
10.86
8.65
10.30
132.00
93.39
94.82
89.96
95.38
233.9
10.84
8.55
10.46
118.25
101.84
103.28
96.78
105.47
246.3
11.04
9.40
10.50
123.00
108.76
110.19
103.47
112.62
268.5
12.19
10.13
11.30
j c/kg
c/kg
c/kg
jc/kg
Jc/kg
L c/kg
c/kg
159.2
252.2
148.9
187.2
191.0
175.4
195.2
195.2
272.4
190.9
203.6
252.2
192.1
166.5
251.0
323.9
246.8
237.0
301.8
187.8
221.2
181.3
267.2
172.8
178.0
226.4
141.5
166.0
200.0
255.1
188.3
199.9
231.1
208.7
159.8
199.9
271.1
200.4
211.0
254.7
211.3
167.1
199.7
296.1
202.1
225.6
296.7
207.1
173.0
247.7
328.5
247.3
234.6
305.2
176.6
221.8
250.4
346.8
254.5
239.7
304.3
173.1
241.7
272.6
330.5
268.8
226.3
306.1
163.2
209.5
260.8
310.2
245.4
244.2
291.6
221.4
219.5
1 S/mt
S/mt
b/S/mt
$Imt
S/mt
S/mt
L$Imt
b/ 5/mt
607
403
970
478
581
209
599
269
919
607
1,352
780
888
307
881
384
1,395
926
2,056
1,161
1.388
450
1,394
562
754
499
1,170
609
678
256
710
318
900
599
1,190
762
876
260
794
338
923
607
1,397
822
917
309
917
396
1,098
724
1,651
928
1,084
402
1,105
485
1,379
914
2.007
1,156
1,375
443
1,384
563
1.382
921
1,958
1,160
1,386
453
1,400
572
1,471
972
2.203
1,249
1,462
443
1,476
575
1,443
963
2.200
1,174
1,428
470
1,425
558
Grains
Barley
$/mt
Maize
5/mt
Rice, Thailand, 5%
L S/mt
Rice, Thailand, 25%
$Imt
Rice, Thailand, 35%
S/mt
*
Rice,Thai, A1.Special / Super S/mt
Sorghum
S/mt
Wheat, Canada
S/mt
Wheat, US, HRW
1 S/mt
WheatUSSRW
S/mt
116.6
121.9
304.9
277.1
272.0
219.5
122.9
216.8
192.0
159.0
172.4
163.7
326.4
306.5
300.1
272.3
162.7
300.4
255.2
238.6
222.1
226.9
585.3
na.
n.a.
522.8
224.1
600.5
399.4
375.1
153.4
170.9
315.6
292.6
285.8
254.5
175.2
232.4
198.4
173.7
167.8
159.4
319.1
297.4
288.8
257.0
151.4
244.7
205.7
187.0
173.8
152.5
327.1
306.5
298.0
265.7
150.7
309.0
274.9
267.5
194.6
171.9
344.0
329.5
327.7
312.0
173.4
415.3
341.9
326.2
216.8
220.4
478.1
na.
n.a.
522.8
218.7
621.7
411.8
384.1
216.4
220.1
464.8
n.a.
na.
432.8
218.5
732.4
425.0
388.7
228.6
234.4
594.0
n.a.
n.a.
537.6
224.9
618.8
439.7
419.6
237.8
246.6
907.0
n.a.
na.
762.7
240.3
537.1
362.2
348.2
Other Food
Bananas EU
Bananas US
Fishmeal
Meat, beef
Meat, chicken
Meat, sheep
Oranges
Shrimp, Mexico
SugarEU domestic
SugarUSdomestic
Sugar, world
897
677
1,166
254.7
138.8
403.6
829
1,024
64.56
48.76
32.59
1,037
676
1,177
260.3
156.7
412.0
957
1,010
68.09
45.77
22.22
1,426
869
1,137
287.1
160.2
462.5
1,136
1,106
75.44
44.93
28.27
1,036
647
1,251
261.2
147.9
394.2
817
988
65.10
45.17
23.49
1,045
705
1,260
259.9
159.3
399.3
893
1,003
66.98
46.47
20.90
999
699
1,123
260.3
163.0
416.5
1,135
1.003
68.28
46.98
21.86
1.068
652
1,075
259.8
156.7
437.9
982
1,045
72.00
44.48
22.61
1.421
836
1,126
282.1
158.8
453.6
1,103
1,103
74.51
44.85
28.42
1,451
792
1,114
283.6
158.7
451.9
1,049
1,105
73.27
44.40
29.78
1,643
1,027
1,161
294.0
162.8
479.4
1,231
1,113
77.14
45.53
29.10
1,440
967
1,168
302.3
164.5
489.3
1,234
1,113
78.22
45.16
27.82
318.5
239.4
595.6
623
749.3
698.6
381.3
268.0
640.7
760
806.3
767.0
537.8
293.0
641.9
1,032
880.5
851.4
371.3
264.8
629.5
720
794.4
727.4
372.8
262.0
639.6
725
807.4
751.3
371.6
269.6
646.7
756
820.9
769.8
409.6
275.5
647.0
838
802.5
819.4
530.8
293.4
640.4
1,036
860.3
850.2
545.7
288.4
639.2
1,083
855.8
849.4
562.8
304.9
640.9
1,069
912.5
856.0
558.9
291.8
646.6
1,021
941.0
855.0
126.7
133.5
231.3
210.8
139.5
142.9
248.0
229.0
167.5
174.3
294.5
277.9
128.8
135.9
241.6
223.3
127.4
130.0
251.6
233.9
148.9
150.3
234.2
213.8
153.0
155.5
264.8
245.1
167.9
174.3
292.6
275.5
165.5
170.0
297.4
281.3
176.8
187.3
298.5
280.9
166.3
174.3
300.3
285.3
Non Energy Commodities
Agriculture
Beverages
Cocoa
Coffee, Arabica
Coffee, robusta
Tea, auctions (3), average
Tea, Colombo auctions
Tea, Kolkata auctions
Tea, Mombasa auctions
Food
Fats and Oils
Coconut oil
Copra
Groundnutoil
Palm oil
Palmkemel oil
Soybean meal
Soybean oil
Soybeans
Raw Materials
Timber
Logs, Cameroon
Logs, Malaysia
Plywood
Sawnwood, Cameroon
Sawnwood, Malaysia
Woodpulp
Other Raw Materials
Cotton A Index
Cotton Memphis
Rubber, US
Rubber, Singapore
S/mt
LS!bbl
$/bbl
a/ $/bbl
1S/bbI
t2ooo,1oo
S/mt
j S/mt
S/mt
L c/kg
j c/kg
c/kg
L S/mt
c/kg
jc/kg
dkg
1c/kg
$/cum
i 5/cum
c/sheets
$/cum
j $/cum
S/mt
t
c/kg
c/kg
c/kg
c/kg
COMMODITY PRICE DATA
Commodity
UnIt
Fertilizers
DAP
Phosphate rock
Potassium chloride
TSP
Urea, E. Europe, bulk
j$/mt
L S/mt
j S/mt
J$/mt
.j S/mt
Metals and Minerals
Aluminum
1 S/mt
Copper
S/mt
$/toz
Gold
c/dmtu
Iron ore
c/kg
Lead
J$/mt
Nickel
c/toz
Silver
**Steel products index Japan j2o00=100
Steelcrcoilsheet.Japan,Reinsj$/mt
Steel hr coilsheet, Japan, Reintj S/mt
Steel,rebar,Japan,ReinstatedS!mt
Steel wire rod, Japan, Reinstatrj S/mt
c/kg
Tin
j c/kg
Zinc
Annual averages
Jan-Dec Jan-Dec Jan-Apr
2006
2007
2008
Quarterly averages
Jan-Mar Apr-Jun Jul-Sep Oct-Dec Jan-Mar
2007
2007
2007
2007
2008
260
44.2
174.5
202
222.9
433
70.9
200.2
339
309.4
267.7
395.2
793
386.0
344
45.5
175.6
226
297.4
2,570
6,722
604.3
77.4
129.0
24,254
1,157
177.6
693.8
600.0
443.8
581.3
878
327.5
2,638
7,118
696.7
84.7
258.0
37,230
1,341
178.1
650.0
550.0
521.5
533.3
1,454
324.2
2,797
8,018
922.5
140.6
288.0
28,908
1.761
226.6
771.9
712.5
677.1
785.5
1,875
238.8
2,801
5,933
650.3
84.7
178.7
41,440
1,332
173.8
650.0
550.0
484.2
500.0
1,273
345.6
945
NEW World Bank commodity price Indices for low and middle
220.9
244.8
Energy
192.1
224.8
Non Energy Commodities
150.4
Agriculture
180.5
145.4
Beverages
169.9
147.0
184.7
Food
137.9
FatsandOils
208.8
149.8
189.0
Grains
156.4
149.0
Other Food
161.4
175.8
RawMaterials
126.0
Timber
136.8
200.0
Other Raw Materials
218.5
168.6
Fertilizers
240.1
280.3
Metals and Minerals
314.0
431
59.9
331
291.3
433
80.0
209.4
375
283.6
522
98.3
230.8
425
365.4
860
234.4
367.7
715
357.6
828
190.0
385.0
729
325.3
1,045
323.1
445.0
876
377.8
1,201
367.5
477.6
1,029
471.3
2.761
7,641
667.4
84.7
217.6
48,055
1,336
175.6
650.0
550.0
540.8
530.0
1,410
366.4
2,546
7,712
681.1
84.7
314.3
30.205
1,273
176.2
650.0
550.0
504.2
550.0
1,498
322.7
2,444
7,188
788.0
84.7
321.5
29,219
1,424
186.7
650.0
550.0
556.8
553.3
1,634
262.3
2.743
7,796
926.8
140.6
289.9
28,957
1,765
221.7
762.5
700.0
639.4
754.0
1,778
243.0
2,777
7,888
922.3
140.6
308.0
27,955
1,767
226.6
800.0
750.0
598.5
725.0
1,721
243.8
3,005
8,439
968.4
140.6
300.9
31,225
1,922
239.6
800.0
750.0
745.0
857.0
1,980
251.1
2,959
8,685
909.7
140.6
282.3
28,763
1,751
241.1
800.0
750.0
790.0
880.0
2,166
226.4
251.1
228.8
183.3
173.3
189.7
216.2
188.3
156.1
172.8
138.9
209.8
240.2
320.8
298.6
237.3
200.9
179.4
212.9
259.1
215.6
149.7
182.8
137.2
232.7
292.0
305.7
331.1
281.4
236.6
210.7
257.2
310.2
274.6
171.9
199.6
146.8
257.3
409.4
358.7
328.4
282.2
238.3
216.4
258.3
313.4
275.3
170.5
200.6
145.7
260.7
384.2
360.8
352.6
301.0
251.8
222.8
276.5
325.5
305.0
186.2
206.4
155.0
262.6
489.9
380.5
376.6
306.3
254.7
213.3
282.5
317.8
340.7
183.5
208.0
157.2
263.6
574.7
381.7
184.8
Income countrles( 2000 =100)
342.4
201.1
228.5
287.6
206.6
226.7
241.1
164.8
173.0
211.3
158.9
167.9
263.5
163.4
172.6
312.1
168.8
191.1
291.1
177.2
174.7
174.8
143.8
146.5
201.7
171.4
176.4
149.4
134.9
136.2
258.9
211.3
220.3
450.7
203.5
224.6
364.4
292.6
337.1
a! Included in the energy index (2000=100)
S
US dollar
mt metric ton
=
0
Monthly averages
Feb
Mar
Apr
2008
2008
2008
b/ Included in the non-energy index (2000=100)
Cl Steel not included in the non-energy index
US cent bbl = barrel cum = cubic meter dmtu = Dry Metric Ton Unit kg = kilogram mmbtu = million British thermal units
toz troy oz n.a. = not available n.q. = no quotation
=
Description of
Price Series
Coal (Australian), thermal, fob. p.ers, Newcastle!Pott Kembla, 6,300 kcalikg (11340 btuflb). less than 0.8%, sulfur 13% ash beginring January 2002;
previously 6.667 kcaVkg (12,000 btuflb). less than 1.0% sulfur, 14% ash
Crude ot (spot), average spot price of Brent, Dubai and West Texas Intermediate, equally weighed
Crude ot (spot), U.K. Brent 38’ API, fob. U.K ports
Crude oti (spot). Dubai Fetch 32’ API, fob Dubai
Crude ci (spot), West Texas Intermediate (Wl’l) 40’ API. f.o.b. k5dland Texas
Natural Gas Index, composite index weighted by coraurnption volumes for Europe, US and Japan Iquefled natural gas (LNG).
Natural Gas (Europe), average Import border price including U.K. for 1991 May, 2000; from June 2000 onwards European Import price excluding U.K
Natural Gas (U.S.), spot price at Henry Hub, Louisiana
Natural gas LNG (Japan). Import price, at, recent two months’ averages are estimates.
-
Cocoa (ICCO), International Cocoa Organation daily price, average of the first three positions on the terminal markets of New York and London,
nearest three future trading months.
Coffee (ICO). International Coffee Organization indicatur price, other mid Arabicas, average New York and BremerVHambLrg markets, ex-dock
Coffee (ICO), International Coffee Orgardzation Indicator price, Robustas, average New York and Le HavrelMarseiies markets, ex-dock
Tea average three auctions, arittrnetic average of quotations at Kolkata, Cotornto and Mornbasalhiairobi.
Tea (Colombo auctions), Sri Laran origIn, at tea, arithmetic average of weekly quotes.
Tea (Kolkata auctions), leaf, include excise duty, arithmetic average of weekly quotes.
Tea (Mombasa,’Nairobl auctions), African origin, al tea, arithmetic average of weekly quotes.
,
Coconut ol (Philippines/Indonesian), bulk, c.I.f. Rotterdam
Copra (PhilippInes/Indonesian), bulk, c.i.f. N.W. Europe
Groundrsit oil (any origin), c.i.f. Rotterdam
Palm oil (Malaysian), 5% bu&, c.l.f. N. W. Europe
Palmkernel Di (Malaysian), c.l.f. Rotterdam
Soybean meal (any origin), Argentine 45/46% exlraction, cit. Rotterdam; prior to 1990. US 44%
Soybean oil (Dutch), crude, f.o.b. ex-mill
Soybeans (US), eLf. Rotterdam
2-
COMMODITY PRICE DATA
Annual averages
Quarterly averages
Jan-Dec Jan-Dec Jan-Apr
Jan-Mar Apr-Jun Jul-Sep Oct-Dec Jan-Mar
Unit
2006
2007
2008
Commodity
2007
2007
2007
2007
2008
Barley (Canadian), feed, Western No. 1, Winnipeg Commodity Exchange, spot, wholesale farmers’ price
Maize (US), no.2, yetow, f.o.b. US Gulf ports
Rice (Thai), 5% broken, white rice (WR), milled, indicative price based on weekly surveys of export transactions, government standard, fob. Bangkok
Rice (ThaI), 25% broken, WR, milled Indicative survey price, government standard, Lob. Bangkok
Rice (Thai), 35% broken, Wit, mlled, indicative survey price, govemment standard, f.o.b. Bangkok
Rice (ThaI), 100% broken, A.l Super from 2006 onwards, broken kernel obtained from the miltng of WR 15%, 20%, and 25%, government standard,
f.o.b. Bangkok; prior to 2006, Al Special also refers to 100% broken rice similar to Al Super, but slightly lower in grade than Al Super.
Sorghum (US), no.2 mile yetow, f.o.b. Gulf ports
Wheat (Canadian), no.1, Western Red Spring (CWRS), in store, St. Lawrence, export price
Wheat (US). no.1, hard red winter, ordinary protein, export price delivered at the Gulf port for prompt or 30 days shipment
Wheat (US). no.2, soft red winter, export price delivered at the Gulf port for prompt or 30 days shipment
Monthly averages
Feb
Mar
Apr
2008
2008
2008
Banaras (Central & South American). major brands, c.i.f. Hamburg
Banares (Central & South American), major brands, US Import price, free on truck (f.o.t.) US Gulf ports
F’ehmeal (any origin), 64-65%, up to December 2003 c&f Hamburg, nis; from January 2004 onwards Breman
Meat, beef (Australian/New ZealandX up to October 2002, cow forequarters, frozen boneinas, 85% chemical lean, c.i.f. U.S. port (East Coast), ex-docic from November,
2002 onwards chucks and cow forequarters
Meat chicken (US), broiler/fryer, whole bIrds, 2-1/2 to 3 pounds. USDA grade A, ice-pecked, Georgia Dock preliminary weighted average, wtrolesale
Meat, sheep (New Zealend frozen whole carcasses Prune Meditsn (PM) beginning January 2006, wholesale, Smithfield, London; prior to January 2006 Prime Light (PL)
Oranges (Mediterranean exporters) navel, EEC indicative unport price, c.i.f. Paris
Shrimp, (Mexican), west coast frozen, white, No.1, shell-on, headtera, 26 to 30 cor,rt per poled, wholesale price at New York
Sugar (EU), European Union negotiated import price for raw unpaciraged sugar from African, Caribbean and Pacific (ACP) under Lome Conventons,
c.l.f. European ports
Sugar (US), import price, nearest future, c.i.f. NewYork
Sugar (world), International Sugar Agreement (ISA) deity price, raw, Lob. end stowed at greater Caribbean ports
Logs (West African), sepele. high quatty (loyal and marchand LM), fob. Douals, Cameroon begtrsiing January 1996, LM 80 centimeter or more
Logs (Malaysian), merenti, Ssrswak, eale price charged by importers, Tokyo; prior to February 1993, average of Ssbsh and Ssrawsk weighted by Japanese
import vokrmes
Plywood (African end Southeest Asian), Leusrt 3-ply, adre, 91 cmx 182cm x4 mm, wtmlesale price, spot Tokyo
Sawrwood (Cemeroonlan), sapele. width 6 inches or more, terrgth 6 feet or more, f.a.a. Cemeroonlan ports
Sswnwood (Malaysian), dark red sereys/meranti. select and better quality, average 7 to 8 inches; length average 12 to 14 inches; thickness 1 to 2 inctr(es);
kiln dry, c. & L UK ports, with 5% agents commission; beginning Jermery 200 indudes premium ($20 to $30) for products of certified sustsfrreble forest
Woodpdp (Swedish), softwood, sulphate, bleached, sir-dry weight c.i.f. North Sea ports
—
Cotton (Cotton Outlook Cottooks, ‘mdcxl, midding 1-3/32 ‘md traded in Far East C/F starting 2006; for
years Northern Europe, c.i.f.
Cotton (US), Memphin/Eeslem, mIddling 1-3/32 inch, c.i.f. Northem Europe. one of the 15 styles based on which the Codook A Index in computed
Rubber (any origin), Ribbed Smoked Sheet (RSS) no. 1 • in bales, Rubber Traders Association (RTA), spot New York
Rubber (Asian), RSS no. 1, in bales, Rubber Association of Singapore Commodity Exchange (RASCE)/ Singapore Commodity Exchange, midday buyers’
asking price for prompt or 30 days delivery prior to June 1992, spot Singapore
DAt’ (diammonkrm phosphate), standard size, bulk, spot fob. US Gui
Phosphate rock (Moroccan), 70% BPL, contract f.a.a. Casablanca
Potassium chloride (muriate of potash), standard grade, spot f.o.b. Vancouver
TSP (triple supemphosphate), upto September 2006 bulk, spot f.o.b. US Gull from October2006 onwerds Tunisian, grenuler, f.o.b.
Urea, (Black See), bulk, spot for 1965-91 (June) fob. Eestem Europe; 1991 (July) onwards f.o.b. Black Sea (primarily Yuzhnw)
Aluminum (LME) London Metal Exchange, unalloyed primary ingota, high grade, minimum 99.7% purity, cash until December 2004; thereafter settlement price
Copper (LME). grade A, minimum 99.9935% purity, cathodes and wire bar shapes, settlement price
Gold (UK). 99.5% fine, London afternoon fixing, average of deity rates
Iron ore (Brezitan), Companhia Vale do Rio Doce (CVRD) Csrejss sinter feed, for years 2005-08, 67.50% Fe (iron) content (dry weight) ores, moisture content 8.0%;
for year 2004, 67.40% Fe; 2000-03, 67.55% Fe, moisture 7.6 8.0 %; contract price to Europe, f.o.b. Ponts da Madeira. Unit dry metric ton unit (dmtu) stands for
mt 1% Fe-unit. To convert price in cants/dmtu to $/dmt SSF (dry ore), multiply by percent Fe content.
Lead (LME), refined, 99.97% purity, settlement price
Nickel (LME), cathodes, minimum 99.8% purity, official morning session, weekly average bid/asked cash until December 2004; thereafter settlement price
Silver (Handy& Harman), 99.9% grade refined, New York
DROPPED due to update difficuttles: Steel products (Belgium), composite price index for eight selected steel products, export, f.o.b., ordinary commercial quality.
REINSTATED Sleet products price index, 2000=100, (Japanese), composite price index for eight selected sleet products besed on quotstiore f.o.b. Japan excluding
shipments to the United States, including China after 2002. weighted by product shares of apparent combined consunpiton (volume of deliveries) at Germany,
Japan and the United Stales. The eight products are as follow. rebar (concrete reinforcing bare). march bar (merchant bars), wire rod, section (H-shape),
plate (medium), hot roted coil/sheet, cold roted cot/sheet and galvanized iron sheet
‘tin (LME), refined, 99.85% purity, settlement price
Zinc (LME). epaulet high grade, minimum 99.995% purity, weekly average bid/asked price, official morning session; prior to April 1990, high grade, minimum 99.95%
ptsity, settlement price
-
* *
Development Prospects Group
Development Economics Vice Presidency
Worid Bank
May 6,2008
1818 H Street, NW. • MSN MC 2-200
Washington, D.C. 20433 U.S.A.
Tel. (202) 473 3862
Emsfl GCMworidbank.org
-
This article presents results from the
first statistically significant study of
cost escalation in transportation in
frastructure projects. Based on a sam
ple of 258 transportation infrastruc
ture projects worth US$90 billion and
representing different project types.
geographical regions, and historical
periods, it is found with overwhelming
statistical significance that the cost es
timates used to decide whether such
projects should be built are highly and
systematically misleadIng. Underesti
mation cannot be explained by error
and is best explained by strategic mis
representation, that is, lying. The pol
icy implications are clear: legislators,
administrators, investors, media rep
resentatives, and members of the pub
lic who value honest numbers should
not trust cost estimates and cost-ben
efIt analyses produced by project pro
moters and their analysts.
Flyvbjerg is a professor of planning with
the Department of Development and Plan
ning. Aalborg University, Denmark. He is
founder and director of the university’s re
search program on transportation infra
structure planning and was twice a Visiting
Fulbright Scholar to the U.S. His latest
books are Rationality and Power (University
of Chicago Press, 1998) and Making Social
Science Matter (Cambridge University Press.
2001). He is currently working on a book
about megaprojects and risk (Cambridge
University Press). Hoim is an assistant pro
fessor of planning with the Department of
Development and Planning. Aalborg Uni
versity, and a research associate with the
university’s research program on transpor
tation infrastructure planning, Her main In
terest is economic appraisal of projects.
Buhi is an associate professor with the De
partment of Mathematics, Aalborg Univer
sity, and an associate statistician with the
university’s research program on transpor
tation infrastructure planning.
Underestimating
Costs in Public
Works Projects
Error or Lie?
Bent Flyvbjerg, Mette Skamris Hoim, and Søren Buhi
C
omparative studies of actual and estimated costs in transportation
infrastructure development are few. Where such studies exist, they are
typically single-case studies or they cover a sample of projects too
small to allow systematic, statistical analyses (Bruzelius et al., 1998; Fouracre
etal., 1990; Hall, 1980; Nijkamp & Ubbels, 1999; Pickrell, 1990; Skamris &
Flyvbjerg, 1997; Szyliowicz & Goetz, 1995; Walmsley & Pickett, 1992). To
our knowledge, only one study exists that, with a sample of 66 transporta
tion projects, approaches a large-sample study and takes a first step toward
valid statistical analysis (Merewitz, 1 973a, 1 973b) ,1 Despite their many mer
its in other respects, these studies have not produced statistically valid an
swers regarding the question of whether one can trust the cost estimates
used by decision makers and investors in deciding whether or not to build
new transportation infrastructure. Because of the small and uneven sam
ples used in existing studies, different studies even point in opposite direc
tions, and researchers consequently disagree regarding the credibility of cost
estimates. Pickrell (1990), for instance, concludes that cost estimates are
highly inaccurate, with actual costs being typically much higher than esti
mated costs, while Nijkamp and Ubbels (1999) claim that cost estimates are
rather correct. Below we will see who is right.
The objective of the study reported here was to answer the following
questions in a statistically valid manner: How common and how large are
differences between actual and estimated costs in transportation infra
structure projects? Are the differences significant? Are they simply random
errors? Or is there a statistical pattern to the differences that suggests other
explanations? What are the implications for policy and decision making re
garding transportation infrastructure development?
lournal of the American Planning Association,
Vol. 68, No.3, Summer 2002. C) American
Planning Association, Chicago, IL,
APA Journal • Surnxner 2002 • Vol. 68, No. 3 279
BENT FLYVBJERG. METE SKAMRIS HOLM. AND SØREN BURL
Four Steps to Understanding
Deceptive Cost Estimation
We see four steps in the evolution of a body of schol
arly research aimed at understanding practices of cost
underestimation and deception in decision making for
transportation infrastructure. The first step was taken
by Pickrell (1990) and Fouracre, A]lport, and Thomson
(1990), who provided sound evidence for a small number
of urban rail projects that substantial cost underestima
tion is a problem, and who implied that such underesti
mation may be caused by deception on the part of pro
ject promoters and forecasters. The second step was
taken by Wachs (1990), who established—again for a
small sample of urban rail projects—that lying, under
stood as intentional deception, is, in fact, an important
cause of cost underestimation. Wachs began the diffi
cult task of charting who does the lying, why it occurs,
what the ethical implications are. etc.
The problem with the research in the first two steps
is that it is based on too few cases to be statistically sig
nificant; the pattern found maybe due to random prop
erties of the small samples involved. This problem is
solved in the third step, taken with the work reported in
this article. Based on a large sample of transportation in
frastructure projects, we show that (1) the pattern of cost
underestimation uncovered by Pickrell and others is of
general import and is statistically significant, and (2) the
pattern holds for different project types, different geo
graphical regions, and different historical periods. We
also show that the large-sample pattern of cost underes
timation uncovered by us lends statistical support to the
conclusions about lying and cost underestimation ar
rived at by Wachs for his small sample.
The fourth and final step in understanding cost un
derestimation and deception would be to do for a large
sample of different transportation infrastructure pro
jects what Wachs did for his small sample of urban rail
projects: establish whether systematic deception actu
ally takes place, who does the deception. why it occurs,
etc. This may be done by having a large number of fore
casters and project promoters, representing a large num
ber of projects, directly express, in interviews or surveys,
their intentions with and reasons for underestimating
costs, This is a key topic for further research.
In sum, then, we do not claim with this article to
have provided final proof that lying is the main cause of
cost underestimation in transportation infrastructure
projects. We claim, however, to have taken one signifi
cant step in a cumulative research process for testing
whether this is the case by establishing the best and
largest set of data about cost underestimation in trans
portation infrastructure planning so far seen, by carry-
280 APA Journal • Summer 2002 • Vol. 68, No. 3
ing out the first statistically significant study of the is
sues involved, and by establishing that our data support
and give statistical significance to theses about lying de
veloped in other research for smaller, statistically non
significant samples.
As part of further developing our understanding of
cost underestimation, it would also be interesting to
study the differences between projects that are approved
on a competitive basis, by voters at an election, and those
that are funded through formula-based allocations. One
may speculate that there is an obvious incentive to make
a project look better, and hence to underestimate costs,
in the campaign leading up to an election. A good sin
gle-case study of this is Kain’s (1990) article about a rail
transit project in Dallas. Votes are cast more often for
large rail, bridge, and tunnel projects than for road pro
jects. For example, most U.S. highway funds are distrib
uted to states based on a formula (i.e., there is no com
petitive process). A state department of transportation
(DOT) is likely to have a fixed annual budget for con
struction. The DOT leadership would presumably want
fairly accurate cost estimates before allocating the bud
get. One may speculate that large cost underestimation
is less likely in this situation. There are exceptions to this
scenario. Sometimes DOT officials want to persuade
state legislators to increase their budget. And states oc
casionally submit bond issue proposals to voters. In Eu
rope, the situation is similar on important points, al
though differences also exist. This may explain the result
found below, that cost underestimation is substantially
lower for roads than for rail, bridges, and tunnels, and
that this is the case both in the U.S. and Europe. Need
less to say, more research is necessary to substantiate this
observation.
Finally, we want to emphasize that although the
project sample used in this study is the largest of its kind,
it is still too small to allow more than a few subdivisions,
if comparative statistical analyses must still be possible.
Therefore, in further work on understanding cost un
derestimation, the sample should be enlarged to better
represent different types of projects and different geo
graphical locations. As to project types, data for more
private projects would be particularly useful in allowing
statistically valid comparisons between public and pri
vate sector projects. Such comparisons do not exist
today, and nobody knows whether private projects per
form better or worse than public ones regarding cost un
derestimation. The sample should also be enlarged to
contain data for more fixed-link and rail projects. Such
data would allow a better (i.e., a statistically corrobo
rated) comparative understanding of cost underestima
tion for more specific subtypes of projects such as
bridges, tunnels, high-speed rail, urban rail, and conven
UNDERESTIMATING COSTS IN PUBLIC WORKS PROJECTS
tional rail. Such an understanding is nonexistent today.
As to geography, immediate rewards would be gained
from data for projects outside Europe and North Amer
ica, especially for fixed links and roads. But even for Eu
rope and North America, data on more projects are
needed to allow better comparative analysis.
Measuring Cost Inaccuracy
The methods used in our study are described in the
Appendix. All costs are construction costs. We follow in
ternational convention and measure the inaccuracy of
cost estimates as so-called “cost escalation” (often also
called “cost overrun”: i.e., actual costs minus estimated
costs in percent of estimated costs). Actual costs are de
fined as real, accounted construction costs determined
at the time of project completion. Estimated costs are de
fined as budgeted, or forecasted, construction costs at the
time of decision to build. Although the project planning
process varies with project type, country, and time, it is
typically possible for any given project to identify a spe
cific point in the process as the time of decision to build.
Usually a cost estimate was available at this point in time
for the decision makers, If not, then the closest available
estimate was used, typically a later estimate resulting in a
conservative bias in our measure for inaccuracy (see the
Appendix).All costs are calculated in fixed prices in Euros
by using the appropriate historical, sectoral, and geo
graphical indices for discounting and the appropriate ex
change rates for conversion between currencies.
Project promoters and their analysts sometimes ob
ject to this way of measuring cost inaccuracy (Flyvbjerg
et al., in press). Various cost estimates are made at dif
ferent stages of the process: project planning, decision
to build, tendering, contracting, and later renegotia
tions. Cost estimates at each successive stage typically
progress toward a smaller number of options, greater de
tail of designs, greater accuracy of quantities, and better
information about unit price. Thus, cost estimates be
come more accurate over time, and the cost estimate at
the time of making the decision to build is far from final.
It is only to be expected, therefore, that such an early es
timate would be highly inaccurate. And this estimate
would be unfair as the basis for assessing the accuracy
of cost forecasting, or so the objection against using the
time-of-decision-to-build estimate goes (Simon, 1991).
We defend this method, however, because when the
focus is on decision making, and hence on the accuracy
of the information available to decision makers, then it
is exactly the cost estimate at the time of making the de
cision to build that is of primary interest. Otherwise it
would be impossible to evaluate whether decisions are
informed or not. Estimates made after the decision to
build are by definition irrelevant to this decision. What
ever the reasons are for cost increases after decision mak
ers give the go-ahead to build a project, or however large
such Increases are, legislators and citizens—or private in
vestors in the case of privately funded projects—are enti
tled to know the uncertainty of budgets. Otherwise
transparency and accountability suffer. We furthermore
observe that if the inaccuracy of early cost estimates were
simply a matter of incomplete information and inher
ent difficulties in predicting a distant future, as project
promoters often say it is, then we would expect inaccu
racies to be random or close to random. Inaccuracies,
however, have a striking and highly interesting bias, as
we will see below.
Another objection to using cost at the time of deci
sion to build as a basis of comparison is that this sup
posedly would entail the classical error of comparing ap
ples and oranges. Projects change over the planning and
implementation process. When, for instance, the physi
cal configuration of the original Los Angeles Blue Line
Light Rail project was altered at substantial cost to com
prise grade-crossing improvements, upgrading of adja
cent streets, better sidewalks, new fences, etc., the pro
ject was no longer the same. It was, instead, a new and
safer project, and comparing the costs of this project
with the costs of the older, less safe one would suppos
edly entail the apples-and-oranges error. A problem with
this argument is that existing research indicates that
project promoters routinely ignore, hide, or otherwise
leave out important project costs and risks in order to
make total costs appear low (Flyvbjerg et al., in press:
Wachs, 1989, 1990). For instance, environmental and
safety concerns may initially be ignored, even though
they will have to be taken into account later in the pro
ject cycle if the project lives on, and the project is more
likely to live on if environmental and safety concerns are
initially ignored. Similarly, ignoring or underplaying
geological risks may be helpful in getting projects ap
proved, and no other risk is more likely to boomerang
back and haunt projects during construction. “Salami
tactics” is the popular name used to describe the prac
tice of introducing project components and risks one
slice at a time in order to make costs appear low as long
as possible. If such tactics are indeed a main mechanism
in cost underestimation, as existing research indicates,
then, clearly, comparing actual project costs with esti
mated costs at the time of decision to build does not en
tail the error of comparing apples and oranges but is
simply a way of tracking how what was said to be a small,
inexpensive apple turned out to actually be a big, expen
sive one,
Finally, we observe that if we were to follow the ob
jections against using the cost estimate at the time of de
APA Journal • Summer 2002 • Vol. 68, No. 3 281
BENT FLYVBJERG. METrE SKAMRIS HOLM, AND SØREN BUHL
cision to build as the basis of tracking cost escalation, it
would be impossible to make meaningful comparisons
of costs because no common standard of comparison
would be available. We also observe that this method Is
the international standard for measuring inaccuracy of
cost estimates (Fouracre et al., 1990; Leavitt et al., 1993;
National Audit Office & Department of Transport,
1992; Nijkamp & Ubbels, 1999; Pickrell, 1990; Walms
ley & Pickett, 1992; World Bank, 1994). This standard
conveniently allows meaningful and consistent com
parisons within individual projects and across projects,
project types, and geographical areas. This standard,
then, is employed below to measure the inaccuracy of
cost estimates in 258 transportation infrastructure pro
jects worth US$90 billion.
Inaccuracy of Cost Estimates
Figure 1 shows a histogram with the distribution of
inaccuracies of cost estimates. If errors in estimating
costs were small, the histogram would be narrowly con
centrated around zero. If errors in overestimating costs
were of the same size and frequency as errors in under
estimating costs, the histogram would be symmetrically
distributed around zero. Neither is the case. We make
the following observations regarding the distribution of
inaccuracies of construction cost estimates:
• Costs are underestimated in almost 9 out of 10
projects. For a randomly selected project, the like
lihood of actual costs being larger than estimated
costs is 86%. The likelihood of actual costs being
lower than or equal to estimated costs is 14%.
• Actual costs are on average 28% higher than
estimated costs (sd=39).
• We reject with overwhelming significance the
thesis that the error of overestimating costs is as
common as the error of underestimating costs
(p.cO.OO1; two-sided test, using the binomial
distribution). Estimated costs are biased, and the
bias is caused by systematic underestimation.
• We reject with overwhelming significance the
thesis that the numerical size of the error of
underestimating costs is the same as the
numerical size of the error of overestimating costs
(p<O.OOl; nonparametric Mann-Whitney test).
Costs are not only underestimated much more
often than they are overestimated or correct, costs
that have been underestimated are also wrong by a
substantially larger margin than costs that have
been overestimated.
We conclude that the error of underestimating costs is
significantly much more common and much larger than
282 APA Journal • Summer 2002 • Vol. 68, No. 3
the error of overestimating costs. Underestimation of
costs at the time of decision to build is the rule rather
than the exception for transportation infrastructure pro
jects. Frequent and substantial cost escalation is the
result.
Cost Underestimation by Project
Type
In this section, we discuss whether different types of
projects perform differently with respect to cost under
estimation. Figure 2 shows histograms with inaccura
cies of cost estimates for each of the following project
types: (1) rail (high-speed; urban: and conventional,
inter-city rail), (2) fixed link (bridges and tunnels), and
(3) road (highways and freeways). Table 1 shows the ex
pected (average) inaccuracy and standard deviation for
each type of project.
Statistical analyses of the data in Table 1 show both
means and standard deviations to be different with a
high level of significance. Rail projects incur the highest
difference between actual and estimated costs, with an
average of no less than 44.7%, followed by fixed-link proj
ects averaging 33.8% and roads at 20.4%. An F-test falsi
fies the null hypothesis at a very high level of statistical
significance that type of project has no effect on per
centage cost escalation (p<O.OO 1). Project type matters.
The substantial and significant differences among proj
ect types indicate that pooling the three types of projects
in statistical analyses, as we did above, is strictly not ap
propriate. Therefore, in the analyses that follow, each
type of project will be considered separately.
Based on the available evidence, we conclude that
rail promoters appear to be particularly prone to cost
underestimation, followed by promoters of fixed-link
projects. Promoters of road projects appear to be rela
tively less inclined to underestimate costs, although ac
tual costs are higher than estimated costs much more
often than not for road projects as well.
Further subdivisions of the sample indicate that
high-speed rail tops the list of cost underestimation, fol
lowed by urban and conventional rail, in that order. Sim
ilarly, cost underestimation appears to be larger for tun
nels than for bridges. These results suggest that the
complexities of technology and geology might have an
effect on cost underestimation. These results are not sta
tistically significant, however. Even if the sample is the
largest of its kind, it is too small to allow repeated sub
divisions and still produce significant results. This prob
lem can be solved only by further data collection from
more projects.
We conclude that the question of whether there are
significant differences in the practice of cost underesti
UNDERESTIMATING COSTS IN PUBLIC WORKS PROJECTS
40
-
TRANSPORTATION INFRASTRUCTURE
PROJECTS (N=255)
U
C
II,
-
I
aI
40
I
I
-40
P
P
I
0
I
40
I
I
I
80
I
I
120
I
160
I
I
I
200
I
I
240
2O
Cost escalation (%)
FIGURE 1. Inaccuracy of cost estimates in 258 transportation infrastructure projects (fixed prices).
TABLE 1. Inaccuracy of transportation project cost estimates by type of project
(fixed prices).
Project type
Number of
cases (N)
Average cost
escalation (%)
Standard
deviation
Level of
significance (p)
Rail
Fixed-link
Road
All projects
58
33
167
258
44.7
33.8
20.4
27.6
38.4
62,4
29.9
38.7
<0.001
<0.004
<0.001
<0.001
APA Journal • Summer 2002 • Vol. 68, No. 3
283
BENT FLYVBJERG, METE SKAMRIS HOLM, AND SØREN BURL
4n-
RAIL PROJECTS (N =58)
cJ
1
;
LL
10
,•
ci’
0.
.
i4( 4
c
cv
‘
&
I1,
0
40
Cnst e
80
1
1ihori
(%)
‘40
2B0
40.
:cj.
FIXED-LINK PROJECTS (N=33)
il9i
LL
-ac
I
-4
I
I
0
eo
120
I
I
I
200
150
240
280
Co5t esckition (%)
50
[lF.
ROAD PROJECTS (N=167)
40
1.-
CI.
0
40
.
80
120
1’0
•
200
Cmt esc&Uon (%)
FIGURE 2. Inaccuracy of cost estimates in rail, fixed-link, and road projects (fixed prices).
284 APA Journal • Summer 2002 • Vol. 68, No. 3
240
20
UNDERESTIMATING COSTS IN PUBLIC WORKS PROJECTS
mation among rail, fixed-link, and road projects must be
answered in the affirmative. The average difference be
tween actual and estimated costs for rail projects is sub
stantially and significantly higher than that for roads,
with fixed-link projects in a statistically nonsignificant
middle position. The average inaccuracy for rail projects
is more than twice that for roads, resulting in average
cost escalations for rail more than double that for roads.
For all three project types, the evidence shows that it is
sound advice for policy and decision makers as well as
investors, bankers, media, and the public to take any
estimate of construction costs with a grain of salt, espe
cially for rail and fixed-link projects.
Cost Underestimation by
Geographical Location
In addition to testing whether cost underestimation
differs for different kinds of projects, we also tested
whether it varies with geographical location among Eu
rope, North America, and “other geographical areas” (a
group of 10 developing nations plus Japan). Table 2
shows the differences between actual and estimated
costs in these three areas for rail, fixed-link, and road pro
jects. There is no indication of statistical interaction be
tween geographical area and type of project. We there
fore consider the effects from these variables on cost
underestimation separately. For all projects, we find that
the difference between geographical areas in terms of un
derestimation is highly significant (p’cO.OO 1). Geography
matters to cost underestimation.
If Europe and North America are compared sepa
rately, which is compulsory for fixed links and roads be
cause no observations exist for these projects in other ge
ographical areas, comparisons can be made by t-tests (as
the standard deviations are rather different, the Welch
version is used). For fixed-link projects, the average dif
ference between actual and estimated costs is 43.4% in
Europe versus 25.7% North America, but the difference
between the two geographical areas is nonsignificant
4 14). Given the limited number of observations and
(p=O.
the large standard deviations for fixed-link projects, we
would need to enlarge the sample with more fixed-link
projects in Europe and North America in order to test
whether the differences might be significant for more
observations. For rail projects, the average difference be
tween actual and estimated costs is 34.2% in Europe ver
sus 40.8% in North America. For road projects, the simi
lar numbers are 22.4% versus 8.4%. Again, the differences
between geographical areas are nonsignificant (p=O.5 10
and p=0.18
, respectively).
4
We conclude, accordingly, that the highly significant
differences we found above for geographical location
come from projects in the “other geographical areas” cat
egory. The average difference between actual and esti
mated costs in this category is a hefty 64.6%.
Have Estimates Improved Over
Time?
In the previous two sections, we saw how cost un
derestimation varies with project type and geography. In
this section, we conclude the statistical analyses by
studying how underestimation has varied over time. We
ask and answer the question of whether project promot
ers and forecasters have become more or less inclined
over time to underestimate the costs of transportation
infrastructure projects. If underestimation were unin
tentional and related to lack of experience or faulty
methods in estimating and forecasting costs, then, a pri
ori, we would expect underestimation to decrease over
time as better methods were developed and more experi
ence gained through the planning and implementation
of more infrastructure projects.
Figure 3 shows a plot of the differences between ac
tual and estimated costs against year of decision to build
TABLE 2. Inaccuracy of transportation project cost estimates by geographical location (fixed prices).
North America
Europe
Project
type
Rail
Fixed-link
Road
All projects
Average
Number
cost
of projects escalation
(N)
(%)
23
15
143
181
34.2
43.4
22.4
25.7
Standard
deviation
25.1
52.0
24.9
28.7
Average
Number
cost
of projects escalation
(N)
(%)
19
18
24
61
40.8
25.7
8.4
23.6
Other geographical areas
Standard
deviation
36.8
70.5
49.4
54.2
Average
Number
cost
of projects escalation
(N)
(%)
16
0
0
16
Standard
deviation
64.6
49.5
—
—
—
—
64.6
49.5
APA Journal • Summer 2002 • Vol. 68, No. 3 285
BENT FLYVBJERG, METrE SKAMRIS HOLM, AND SØREN BUHL
for the 111 projects in the sample for which these data
are available. The diagram does not seem to indicate an
effect from time on cost underestimation. Statistical
analyses corroborate this impression. The null hypothe
sis that year of decision has no effect on the difference
between actual and estimated costs cannot be rejected
, F-test). A test using year of completion instead
22
(p=O.
of year of decision (with data for 246 projects) gives a
similar result (p=O.
8, F-test).
2
We therefore conclude that cost underestimation
has not decreased over time. Underestimation today is
in the same order of magnitude as it was 10, 30, and 70
years ago. If techniques and skills for estimating and
forecasting costs of transportation infrastructure pro
jects have improved over time, this does not show in the
data. No learning seems to take place in this important
and highly costly sector of public and private decision
making. This seems strange and invites speculation that
the persistent existence over time, location, and project
type of significant and widespread cost underestimation
is a sign that an equilibrium has been reached: Strong
incentives and weak disincentives for underestimation
may have taught project promoters what there is to
learn, namely, that cost underestimation pays off. If this
is the case, underestimation must be expected and it
must be expected to be intentional. We examine such
speculation below. Before doing so, we compare cost un
derestimation in transportation projects with that in
other projects.
Cost Underestimation in Other
Infrastructure Projects
In addition to cost data for transportation infra
structure projects, we have reviewed cost data for several
hundred other projects including power plants, dams,
water distribution, oil and gas extraction, information
technology systems, aerospace systems, and weapons
systems (Arditi et al., 1985; Blake et al., 1976; Canaday,
1980; Department of Energy Study Group, 1975;
Dlakwa & Culpin, 1990; Fraser, 1990; Hall, 1980; Healey,
1964; Henderson, 1977; Hufschmidt & Germ, 1970;
Merewitz, 1973b; Merrow, 1988; Morris & Hough, 1987;
World Bank, 1994, n.d.). The data indicate that other
types of projects are at least as, if not more, prone to cost
underestimation as are transportation infrastructure
projects.
Among the more spectacular examples of cost un
derestimation are the Sydney Opera House, with actual
costs approximately 15 times higher than those pro
jected, and the Concorde supersonic airplane, with a cost
12 times higher than predicted (Hall, nd., p. 3). The data
also indicate that cost underestimations for other proj
286 APA Journal • Summer 2002 • Vol. 68, No. 3
ects have neither increased nor decreased historically,
and that underestimation is common in both First- and
Third-World countries. When the Suez canal was com
pleted in 1869, actual construction costs were 20 times
higher than the earliest estimated costs and 3 times
higher than the cost estimate for the year before con
struction began. The Panama Canal, which was com
pleted in 1914, had cost escalations in the range of 70 to
200% (Summers, 1967, p. 148).
In sum, the phenomena of cost underestimation
and escalation appear to be characteristic not only of
transportation projects but of other types of infrastruc
ture projects as well.
Explanations of Underestimation:
Error or Lie?
Explanations of cost underestimation come in four
types: technical, economic, psychological, and political.
In this section, we examine which explanations best fit
our data.
Tethnical Explanations
Most studies that compare actual and estimated
costs of infrastructure projects explain what they call
“forecasting errors” in technical terms, such as imperfect
techniques, inadequate data, honest mistakes, inherent
problems in predicting the future, lack of experience on
the part of forecasters, etc. (Ascher, 1978; Flyvbjerg et al.,
in press; Morris & Rough, 1987; Wachs, 1990). Few
would dispute that such factors may be important
sources of uncertainty and may result in misleading fore
casts. And for small-sample studies, which are typical of
this research field, technical explanations have gained
credence because samples have been too small to allow
tests by statistical methods. However, the data and tests
presented above, which come from the first large-sam
ple study in the field, lead us to reject technical explana
tions of forecasting errors. Such explanations simply do
not fit the data.
First, if misleading forecasts were truly caused by
technical inadequacies, simple mistakes, and inherent
problems with predicting the future, we would expect a
less biased distribution of errors in cost estimates
around zero. In fact, we have found with overwhelming
) that the distribution of
1
statistical significance (p<O.OO
such errors has a nonzero mean. Second, if imperfect
techniques, inadequate data, and lack of experience were
main explanations of the underestimations, we would
expect an improvement in forecasting accuracy over
time, since errors and their sources would be recognized
and addressed through the refinement of data collection,
forecasting methods, etc. Substantial resources have
UNDERESTIMATING COSTS IN PUBLIC WORKS PROJECTS
300
a
0
200
-
o
0
C
0
to
0
C
o
100
$
-
00
0
0
0
00
0
C-,
0
0
0
0
0
C
-1001910
1920
1930
1940
1950
1960
1970
1950
1990
2000
Year of decision Ia build
FIGURE 3. Inaccuracy of cost estimates in transportation projects over time, 1910—1998 (fixed prices, 111
projects).
been spent over several decades on improving data and
methods. Still our data show that this has had no effect
on the accuracy of forecasts. Technical factors, therefore,
do not appear to explain the data. It is not so-called fore
casting “errors” or cost “escalation” or their causes that
need explaining. It is the fact that in 9 out of 10 cases,
costs are underestimated.
We may agree with proponents of technical expla
nations that it is, for example, Impossible to predict for
the individual project exactly which geological, environ
mental, or safety problems will appear and make costs
soar. But we maintain that it is possible to predict the
risk, based on experience from other projects, thatsome
such problems will haunt a project and how this will af
fect costs. We also maintain that such risk can and
should be accounted for in forecasts of costs, but typi
cally is not. For technical explanations to be valid, they
would have to explain why forecasts are so consistent in
ignoring cost risks over time, location, and project type.
Economic Explanations
Economic explanations conceive of cost underesti
mation in terms of economic rationality. Two types of
economic explanation exist; one explains in terms of eco
APA Journal • Summer 2002 • Vol. 68, No. 3
287
BENT FLYVBJERG, METE SKAMRIS HOLM, AND SØREN BURL
nomic self-interest, the other in terms of the public in
terest. As regards self-interest, when a project goes for
ward, it creates work for engineers and construction
firms, and many stakeholders make money. If these
stakeholders are involved in or indirectly influence the
forecasting process, then this may influence outcomes
in ways that make it more likely that the project will be
built. Having costs underestimated and benefits overes
timated would be economically rational for such stakeholders because it would increase the likelihood of reve
nues and profits. Economic self-interest also exists at the
level of cities and states. Here, too, it may explain cost
underestimation. Pickrell (1990, 1992) pointed out that
transit capital investment projects in the U.S. compete
for discretionary grants from a limited federal budget
each year, This creates an incentive for cities to make
their projects look better, or else some other city may get
the money.
As regards the public interest, project promoters and
forecasters may deliberately underestimate costs in order
to provide public officials with an incentive to cut costs
and thereby to save the public’s money. According to this
type of explanation, higher cost estimates would be an
incentive for wasteful contractors to spend more of the
taxpayer’s money. Empirical studies have identified pro
moters and forecasters who say they underestimate costs
in this manner and with this purpose (i.e., to save public
money; Wachs, 1990). The argument has also been
adopted by scholars, for instance Merewitz (1 973b),who
explicitly concludes that “keeping costs low is more im
portant than estimating costs correctly” (p. 280).
Both types of economic explanation account well for
the systematic underestimation of costs found in our
data. Both depict such underestimation as deliberate,
and as economically rational. If we now define a lie in the
conventional fashion as making a statement intended to
deceive others (Bok, 1979, p. 14; Cliffe et al., 2000, p. 3),
we see that deliberate cost underestimation is lying, and
we arrive at one of the most basic explanations of lying,
and of cost underestimation, that exists: Lying pays off,
or at least economic agents believe it does. Moreover, if
such lying is done for the public good (e.g., to save tax
payers’ money), political theory would classify it in that
special category of lying called the “noble lie,” the lie mo
tivated by altruism. According to Bok (1979), this is the
“most dangerous body of deceit of all” (p. 175).
In the case of cost underestimation in public works
projects, proponents of the noble lie overlook an impor
tant fact: Their core argument—that taxpayers’ money is
saved by cost underestimation—is seriously flawed. Any
one with even the slightest trust in cost-benefit analysis
and welfare economics must reject this argument. Un
derestimating the costs of a given project leads to a
288 APA Journal • Summer 2002 • Vol. 68, No. 3
falsely high benefit-cost ratio for that project, which in
turn leads to two problems. First, the project may be
started despite the fact that it is not economically viable.
Or, second, it may be started instead of another project
that would have yielded higher returns had the actual
costs of both projects been known. Both cases result in
the inefficient use of resources and therefore in waste of
taxpayers’ money. Thus, for reasons of economic effi
ciency alone, the argument that cost underestimation
saves money must be rejected; underestimation is more
likely to result in waste of taxpayers’ money. But the ar
gument must also be rejected for ethical and legal rea
sons. In most democracies, for project promoters and
forecasters to deliberately misinform legislators, admin
istrators, bankers, the public, and the media would not
only be considered unethical but in some instances also
illegal, for instance where civil servants would misinform
cabinet members or cabinet members would misinform
the parliament. There is a formal “obligation to truth”
built into most democratic constitutions on this point.
This obligation would be violated by deliberate under
estimation of costs, whatever the reasons may be. Hence,
even though economic explanations fit the data and help
us understand important aspects of cost underestima
tion, such explanations cannot be used to justify it.
Psychological Explanations
Psychological explanations attempt to explain bi
ases in forecasts by a bias in the mental makeup of proj
ect promoters and forecasters. Politicians may have a
“monument complex,” engineers like to build things,
and local transportation officials sometimes have the
mentality of empire builders. The most common psy
chological explanation is probably “appraisal opti
mism.” According to this explanation, promoters and
forecasters are held to be overly optimistic about project
outcomes in the appraisal phase, when projects are
planned and decided (Fouracreetal., 1990, p. 10; Mackie
& Preston, 1998; Walmsley & Pickett, 1992, p. 11; World
Bank, 1994, p. 86). An optimistic cost estimate is clearly
a low one. The existence of appraisal optimism in pro
moters and forecasters would result in actual costs being
higher than estimated costs. Consequently, the existence
of appraisal optimism would be able to account, in
whole or in part, for the peculiar bias of cost estimates
found in our data, where costs are systematically under
estimated. Such optimism, and associated cost under
estimation, would not be lying, needless to say, because
the deception involved is self-deception and therefore
not deliberate. Cost underestimation would be error ac
cording to this explanation.
There is a problem with psychological explanations,
however. Appraisal optimism would be an important
UNDERESTIMATING COSTS IN PUBLIC WORKS PROJECTS
and credible explanation of underestimated costs if esti
mates were produced by inexperienced promoters and
forecasters, i.e., persons who were estimating costs for
the first or second time and who were thus unknowing
about the realities of infrastructure building and were
not drawing on the knowledge and skills of more expe
rienced colleagues. Such situations may exist and may
explain individual cases of cost underestimation. But
given the fact that the human psyche is distinguished by
a significant ability to learn from experience, it seems un
likely that promoters and forecasters would continue to
make the same mistakes decade after decade instead of
learning from their actions. It seems even more unlikely
that a whole profession of forecasters and promoters
would collectively be subject to such a bias and would
not learn over time. Learning would result in the reduc
tion, if not elimination, of appraisal optimism, which
would then result in cost estimates becoming more ac
curate over time. But our data clearly shows that this has
not happened.
The profession of forecasters would indeed have to
be an optimistic group to keep their appraisal optimism
throughout the 70-year period our study covers and not
learn that they were deceiving themselves and others by
underestimating costs. This would account for the data,
but is not a credible explanation. As observed elsewhere,
the incentive to publish andjustify optimistic estimates
is very strong, and the penalties for having been overop
timistic are generally insignificant (Davidson & Huot,
1989, p. 137; Flyvbjerg et al., in press). This is a better ex
planation of the pervasive existence of optimistic esti
mates than an inherent bias for optimism in the psyche
of promoters and forecasters. And “optimism” calcu
lated on the basis of incentives is not optimism, of
course; it is deliberate deception. Therefore, on the basis
of our data, we reject appraisal optimism as a primary
cause of cost underestimation.
Political Explanations
Political explanations construe cost underestima
tion in terms of interests and power (Flyvbjerg, 1998).
Surprisingly little work has been done that explains the
pattern of misleading forecasts in such terms (Wachs,
1990, p. 145). A key question for political explanations is
whether forecasts are intentionally biased to serve the in
terests of project promoters in getting projects started.
This question again raises the difficult issue of lying.
Questions of lying are notoriously hard to answer, be
cause in order to establish whether lying has taken place,
one must know the intentions of actors. For legal, eco
nomic, moral, and other reasons, if promoters and fore
casters have intentionally fabricated a deceptive cost
estimate for a project to get it started, they are unlikely to
tell researchers or others that this is the case (Flyvbjerg,
1996; Wachs, 1989).
When Eurotunnel, the private company that owns
the tunnel under the English Channel, went public
in 1987 to raise funds for the project, investors were
told that building the tunnel would be relatively straight
forward. Regarding risks of cost escalation, the prospec
tus read:
Whilst the undertaking of a tunneling project of
this nature necessarily involves certain construc
tion risks, the techniques to be used are well
proven.
The Directors, having consulted the
Maitre d’Oeuvre, believe that 10%. would be a
reasonable allowance for the possible impact of un
foreseen circumstances on construction costs.
2
(‘Under Water,’ 1989, p.37)
.
.
.
.
.
Two hundred banks communicated these figures for
cost and risk to investors, including a large number of
small investors. As observed by The Economist (“Under
Water,” 1989), anyone persuaded in this way to buy
shares in Eurotunnel in the belief that the cost estimate
was the mean of possible outcomes was, in effect, de
ceived. The cost estimate of the prospectus was a best
possible outcome, and the deception consisted in mak
ing investors believe in the highly unlikely assumption—
disproved in one major construction project after an
other—that everything would go according to plan, with
no delays; no changes in safety and environmental per
formance specifications; no management problems; no
problems with contractual arrangements, new tech
nologies, or geology; no major conflicts; no political
promises not kept; etc. The assumptions were, in other
words, those of an ideal world. The real risks of cost es
calation for the Channel tunnel were many times higher
than those communicated to potential investors, as evi
denced by the fact that once built, the real costs of the
project were higher by a factor of two compared with
forecasts.
Flyvbj,erg, Bruzelius, and Rothengatter (in press)
document for a large number of projects that the Every
thing-Goes-According-to-Plan type of deception used
for the Channel tunnel is common. Such deception is,
in fact, so widespread that in a report on infrastructure
and development, the World Bank (1994, pp. ii, 22)
found reason to coin a special term for it: the “EGAP
principle.” Cost estimation following the EGAP-princi
pie simply disregards the risk of cost escalation result
ing from delays, accidents, project changes, etc. This is a
major problem in project development and appraisal, ac
cording to the World Bank.
It is one thing, however, to point out that investors,
public or private, were deceived in particular cases. It is
APA Journal • Summer 2002 • Vol. 68, No. 3 289
BENT FLYVBJERG. MEUE SKAMRIS HOLM, AND SØREN BUHL
quite another to get those involved in the deceptions to
talk about this and to possibly admit that deception was
intentional, i.e., that it was lying. We are aware of only
one study that actually succeeded in getting those in
volved in underestimating costs to talk about such is
sues (Wachs, 1986, 1989, 1990). Wachs interviewed pub
lic officials, consultants, and planners who had been
involved in transit planning cases in the U.S. He found
that a pattern of highly misleading forecasts of costs and
patronage could not be explained by technical issues and
were best explained by lying. In case after case, planners,
engineers, and economists told Wachs that they had had
to ‘cook” forecasts in order to produce numbers that
would satisfy their superiors and get projects started,
whether or not the numbers could be justified on tech
nical grounds (Wachs, 1990, p. 144). One typical plan
ner admitted that he had repeatedly adjusted the cost
figures for a certain project downward and the patronage
figures upward to satisfy a local elected official who
wanted to maximize the chances of getting the project
in question started. Wachs’ work is unusually penetrat
ing for a work on forecasting. But again, it is small-sam
ple research, and Wachs acknowledges that most of his
evidence is circumstantial (Wachs, 1986, p. 28). The evi
dence does not allow conclusions regarding the project
population. Nevertheless, based on the strong pattern
of misrepresentation and lying found in his case stud
ies, Wachs goes on to hypothesize that the type of abuse
he has uncovered is “nearly universal” (1990, p. 146;
1986, p. 28) and that it takes place not only in transit
planning but also in other sectors of the economy where
forecasting routinely plays an important role in policy
debates.
Our data give support to Wachs’ claim. The pattern
of highly underestimated costs is found not only in the
small sample of projects Wachs studied; the pattern is
statistically significant and holds for the project popu
lation mean (i.e., for the majority of transportation in
frastructure projects). However, on one point, Wachs
(1986) seems to draw a conclusion somewhat stronger
than is warranted. ‘[F] orecasted costs always seem to be
lower than actual costs” (p. 24) he says (emphasis in orig
inal). Our data show that although “always” (100%) may
cover the small sample of projects Wachs chose to study,
when the sample is enlarged by a factor of 20—30 to a
more representative one, “only” in 86% of all cases are
forecasted costs lower than actual costs. Such trifles—14
percentage points—apart, the pattern identified by Wachs
is a general one, and his explanation of cost underesti
mation in terms of lying to get projects started fit our
data particularly well. Of the existing explanations of
cost development in transportation infrastructure pro
jects, we therefore opt for political and economic expla
290 APA Journal • Summer 2002 • Vol. 68, No. 3
nations. The use of deception and lying as tactics in
power struggles aimed at getting projects started and at
making a profit appear to best explain why costs are
highly and systematically underestimated in transpor
tation infrastructure projects.
Summary and Conclusions
The main findings from the study reported in this
article—all highly significant and most likely conserva
tive—are as follows:
• In 9 out of 10 transportation Infrastructure
projects, costs are underestimated.
• For rail projects, actual costs are on average 45%
higher than estimated costs (sd=38).
• For fixed-link projects (tunnels and bridges),
actual costs are on average 34% higher than
estimated costs (sd=62).
• For road projects, actual costs are on average 20%
higher than estimated costs (sd=30).
• For all project types, actual costs are on average
28% higher than estimated costs (sd=39).
• Cost underestimation exists across 20 nations
and 5 continents; it appears to be a global
phenomenon.
• Cost underestimation appears to be more
pronounced in developing nations than in
North America and Europe (data for rail projects
only).
• Cost underestimation has not decreased over the
past 70 years. No learning that would improve
cost estimate accuracy seems to take place.
• Cost underestimation cannot be explained by
error and seems to be best explained by strategic
misrepresentation, i.e., lying.
• Transportation infrastructure projects do not
appear to be more prone to cost underestimation
than are other types of large projects.
We conclude that the cost estimates used in public
debates, media coverage, and decision making for trans
portation infrastructure development are highly, sys
tematically, and significantly deceptive. So are the costbenefit analyses into which cost estimates are routinely
fed to calculate the viability and ranking of projects. The
misrepresentation of costs is likely to lead to the misal
location of scarce resources, which, in turn, will produce
losers among those financing and using infrastructure,
be they taxpayers or private investors.
We emphasize that these conclusions should not be
interpreted as an attack on public (vs. private) spending
on infrastructure, since the data are insufficient to de
cide whether private projects perform better or worse
UNDERESTIMATING COSTS IN PUBLIC WORKS PROJECTS
than public ones regarding cost underestimation. Nor
do the conclusions warrant an attack on spending on
transportation vs. spending on other projects, since
other projects appear to be as liable to cost underesti
mation and escalation as are transportation projects.
With transportation projects as an in-depth case study,
the conclusions simply establish that significant cost un
derestimation is a widespread practice in project devel
opment and implementation, and that this practice
forms a substantial barrier to the effective allocation of
scarce resources for building important infrastructure.
The key policy implication for this consequential
and highly expensive field of public policy is that those
legislators, administrators, bankers, media representa
tives, and members of the public who value honest num
bers should not trust the cost estimates presented by
infrastructure promoters and forecasters. Another im
portant implication is that institutional checks and bal
ances—including financial, professional, or even crimi
nal penalties for consistent or foreseeable estimation
errors—should be developed to ensure the production
of less deceptive cost estimates. The work of designing
such checks and balances has been begun elsewhere,
with a focus on four basic instruments of accountability:
(1) increased transparency, (2) the use of performance
specifications, (3) explicit formulation of the regulatory
regimes that apply to project development and imple
mentation, and (4) the involvement of private risk capi
tal, even in public projects (Bruzelius et al., 1998; Flyv
bjerg et al., in press).
ACKNOWLEDGMENTS
The authors wish to thank Martin Wachs, Don Pickrell, and
three anonymous JAPA referees for valuable comments on an
earlier draft of the article. Research for the article was sup
ported by the Danish Transport Council and Aalborg Univer
sity, Denmark.
NOTES
1. Merewitz’s (1 973a, 1 973b) study compared cost overrun
in urban rapid transit projects, especially the San Fran
cisco Bay Area Rapid Transit (BART) system, with over
run in other types of public works projects. Merewltz’s
aims were thus different from ours, and his sample of
transportation projects was substantially smaller: 17
rapid transit projects and 49 hIghway projects, compared
with our 58 rail projects, 167 highway projects, and 33
bridge or tunnel projects. In addition to issues of a small
sample, in our attempt to replicate Merewitz’s analysis we
found that his handling of data raises a number of other
issues. First, Merewitz did not correct his cost data for in
flation, i.e., current prices were used instead of fixed ones.
This is known to be a major source of error due to varying
inflation rates between projects and varying duration of
construction periods. Second, in statistical tests, Mere
witz compared the mean cost overrun of subgroups of
projects (e.g., rapid transit) with the grand mean of over
run for all projects, thus making the error of comparing
projects with themselves. Subgroups should be tested di
rectly against other subgroups in deciding whether they
differ at all and, if so, which ones differ. Third, Merewitz’s
two reports (1973a, 1973b) are inconsistent. One (1973a)
calculates the grand mean of cost overrun as the average
of means for subgroups: that is. the grand mean is un
weighted, where common practice is to use the weighted
mean, as appears to be the approach taken in the other
(1973b). Fourth, due to insufficient information, the p
values calculated by Merewitz are difficult to verify; most
likely they are flawed, however, and Merewitz’s one-sided
p-values are misleading. Finally, Merewitz used a debat
able assumption about symmetry, which has more impact
for the nonparametric test used than nonnormality has
for parametric methods. Despite these shortcomings, the
approach taken in Merewitz’s study was innovative for its
time and in principle pointed in the right direction re
garding how to analyze cost escalation in public works
projects. The study cannot be said to be a true large-sam
ple study for transportation infrastructure, however, and
its statistical significance is unclear.
2. The Maitre d’Oeuvre was an organization established to
monitor project planning and implementation for the
Channel tunnel. It was established in 1985. and until 1988
it represented the owners. In 1988 it was reverted to an
impartial position (Major Projects Association, 1994, pp.
151—153).
REFERENCES
Arditi, D., Akan, C. T., & Gurdamar, S. (1985). Cost overruns in
public projects. InternationalJournaiofProject Management.
3(4), 2 18—225.
Ascher, W. (1978). Forecasting: An appraisal forpollcy-makers and
planners. Baltimore: Johns Hopkins University Press.
Blake, C., Cox, D., & Fraize, W. (1976). Analysis ofpivjected vs. ac
tual costs for nuclear and coal-fired power plants (Report re
pared for the United States Energy Research and Devel
opment Administration). McLean, VA: Mitre Corporation.
Bok, S. (1979). Lying: Moral choice in public and private life. New
York: Vintage.
Bruzelius, N., Flyvbjerg, B., & Rothengatter, W. (1998). Big de
cisions, big risks: Improving accountability in mega pro
jects. International Review ofAdministrative Sciences, 64(3),
423—440.
Canaday, H. T. (1980). Construction costoverrunsin electric utilities:
Some trends and implications (Occasional Paper No. 3).
Columbus: National Regulatory Research Institute, Ohio
State University.
Cliffe, L., Ramsey, M., & Bartlett, D. (2000). The politics of lying:
Implications for democracy. London: Macmillan.
APA Journal • Summer 2002 • Vol. 68, No. 3
291
BENT FLYVBJERG, METE SKAMRIS HOLM. AND SØREN BUHL
Commission of the European Union. (1993). Growth, cornpeti
tiveness, employment: The challenges and ways forward into the
21st century (White Paper). Brussels: Author.
Davidson, F. P., & Huot, J.-C. (1989). Management trends for
major projects. ProjectAppraisal, 4(3), 133—142.
Department of Energy Study Group. (1975). North Sea costs es
calation study (Energy Paper No. 8). London: Department
of Energy.
Dlakwa, M. M., & Culpin, M. F. (1990). Reasons for overrun in
public sector construction projects in Nigeria. Interna
tionalJournal ofProjectManagement, 8(4), 2 37—240.
Flyvbjerg, B. (1996). The dark side of planning: Rationality and
Realrationalitht. In S. Mandelbaum, L. Mazza, & R. Bur
chell (Eds.), Explorations In planning theory (pp. 383-394).
New Brunswick, NJ: Center for Urban Policy Research
Press.
Flyvbjerg, B. (1998). Rationality andpower:Democracyinpractice.
Chicago: University of Chicago Press.
Flyvbjerg, B., Bruzelius, N., & Rothengatter, W. (in press).
Megaprojects and risk: An anatomy ofambition. Cambridge,
UK: Cambridge University Press.
Fouracre, P. R., Allport, R. J., & Thomson, J. M. (1990). The per
formance and Impact ofrail mass transit in developing countries
(‘TRRL Research Report 278). Crowthorne, UK: Transport
and Road Research Laboratory.
Fraser, R. M. (1990). Compensation for extra preliminary
and general (P & C) costs arising from delays, variations
and disruptions: The palmiet pumped storage scheme.
Tunneling and Underground Space Technology, 5(3), 205216.
Hall, P. (1980). Great planning disasters. Harmondsworth, UK:
Penguin Books.
Hall, P. (n.d). Great planning disasters revisited. Unpublished
manuscript, Bartlett School, University College, London.
Healey, J. M. (1964). Errors in project cost estimates. Indian Eco
nomic Journal, 12(1), 44—52.
Henderson, P. D. (1977). Two British errors: Their probable
size and some possible lessons. Oxford Economic Papers,
29(2), 159—205.
Hoim, M. K. S. (1999). Inaccuracy of traffic forecasts and cost esti
mates in Swedish road and rail projects. Unpublished manu
script, Aalborg University, Department of Development
and Planning.
Hufschmidt, M. M., & Germ, J. (1970). Systematic errors in
cost estimates for public investment projects. In J. Mar
golis (Ed.), The analysis ofpublic output (pp. 267—315). New
York: Columbia University Press.
Kain, J. F. (1990). Deception in Dallas: Strategic misrepresen
tation in rail transit promotion and evaluation. Journal of
the Amen can PlanningAssociation, 56(2), 184—196.
Leavitt, D., Ennis, S., & McGovern, P. (1993). The cost escalation
ofrail projects: Using previous experience to re-evaluate the calspeed estimates (Working Paper No. 567). Berkeley: Insti
tute of Urban and Regional Development, University of
California.
Lewis, H. (1986). The metro report: The impact ofmetro andpublic
transport integration in Tyne and Wear. Newcastle, UK: Tyne
and Wear Passenger Transport Executive.
292 APA Journal • Summer 2002 • Vol. 68, No. 3
Mackie, P., & Preston, J. (1998). Twenty-one sources of error
and bias in transport project appraisal. Transport Policy,
5(1), 1—7.
Major Projects Association. (1994). Beyond 2000:A source book
formajorprojects. Oxford, UK: Author.
Merewitz, L. (1 973a). I-low do urban rapid ti-ansitprojects compare
in cost estimate experience? (Reprint No. 104). Berkeley: In
stitute of Urban and Regional Development, University
of California.
Merewitz, L. (l973b). Cost overruns in public works. In W.
Niskanen, A. C. Hansen, R. H. Havemann, R. Turvey, & R.
Zeckhauser (Eds.), Benefit cost and policy analysis (pp. 277—
295). Chicago: Aldine.
Merrow, E. W. (1988). Understanding the outcomes otmegaprojects:
A quantitative analysis of very large civilian projects. Santa
Monica, CA: RAND Corporation.
Morris, P. W. C., & Hough, C. H. (1987). The anatomy ofmajor
projects:A study ofthe realityofprojectmanagement. New York:
John Wiley and Sons.
National Audit Office, Department of Transport. (1985). Ex
penditure on motorways and trunk roads. London: National
Audit Office.
National Audit Office, Department of Transport. (1992). Con
ti-a cting for roads. London: National Audit Office.
National Audit Office, Department of Transport: Scottish De
velopment Department: & Welsh Office. (1988). Road
planning. London: Her Majesty’s Stationary Office.
Nijkamp, P., & Ubbels, B. (1999). How reliable are estimates of
infrastructure costs? A comparative analysis. International
Journal of TransportEconomics, 26(1), 23—53.
Pickrell. D. H. (1990). Urban rail transit projects: Forecast versus
actualrldership andcost. Washington, DC: U.S. Department
of Transportation.
Pickrell, D. H. (1992). A desire named streetcar: Fantasy and
fact in rail transit planning. Journal of the American Plan
ningAssociation, 58(2), 158—176.
Riksrevisionsverket. (1994). Infrastrukturinvesteringar: En kost
nadsjamfbrelse mellan plan och utfall i 15 stOrre projekt mom
Vagverketoch Banverket. Stockholm: Author.
Simon, J. (1991). Let’s make forecast and actual comparisons
fair. TRNews, 156, 6—9.
Skamris, M. K., & Flyvbjerg, B. (1997). Inaccuracy of traffic
forecasts and cost estimates on large transport projects.
Transport Policy, 4(3), 14 1—146.
Summers, R. (1967). Cost estimates as predictors of actual
costs: A statistical study of military developments. In T.
Marschak, T. K. Glennan, & R Summers (Eds.), Strategy
for R&D: Studies in the microeconomics of development (pp.
140—189). Berlin: Springer-Verlag.
Szyliowicz, J. S., & Goetz, A. R. (1995). Getting realistic about
megaproject planning: The case of the new Denver Inter
national Airport. PolicySciences, 28(4), 347—367.
Underwater, over budget. (1989, October 7). The Economist, pp.
37—38.
Vejdirektoratet. (1995). Notat om anlgsregnskaber. Copen
hagen: Danish Road Directorate.
Wachs, M. (1986). Technique vs. advocacy in forecasting: A
study of rail rapid transit. Urban Resources, 4(1), 23—30.
UNDERESTIMATING COSTS IN PUBLIC WORKS PROJECTS
Wachs, M. (1989). When planners lie with numbers. Journal of
theAmerican PlanningAssociation, 55(4), 476—479.
Wachs, M. (1990). Ethics and advocacy in forecasting for public policy. Business and Professional Ethics Journal, 9(1—2),
141—157.
Walnisley, D. A., & Pickett, M. W. (1992). The cost and patronage ofrapid transit systems compared with forecasts (Research
Report 352). Crowthorne, UK: Transport Research
Laboratory.
World Bank. (1994). World development report 1994: Infrastruc
ture for development. Oxford, UK: Oxford University Press.
World Bank. (n. d.). Economic analysis ofprojects: Towards a re
suits-oriented approach to evaluation (ECON Report). Washington, DC: Author.
APPENDIX
The first task of the research reported in this paper
was to establish a sample of infrastructure projects sub
stantially larger than what is common in this area of re
search, a sample large enough to allow statistical analy
ses of costs. Here a first problem was that data on actual
costs in transportation infrastructure projects are rela
tively difficult to come by. One reason is that it is quite
time consuming to produce such data. For public sector
projects, funding and accounting procedures are typi
cally unfit for keeping track of the multiple and complex
changes that occur in total project costs over time. For
large projects, the relevant time frame may cover 5, 10, or
more fiscal years from decision to build, until construc
tion starts, until the project is completed and operations
begin. Reconstructing the actual total costs of a public
project, therefore, typically entails long and difficult ar
chival work and complex accounting. For private pro
jects, even if funding and accounting practices may be
more conducive to producing data on actual total costs,
such data are often classified to keep them from the
hands of competitors. Unfortunately, this also tends to
keep data from the hands of scholars. And for both pub
lic and private projects, data on actual costs may be held
back by project owners because more often than not, ac
tual costs reveal substantial cost escalation, and cost es
calation is normally considered somewhat of an embar
rassment to promoters and owners. In sum, establishing
reliable data on actual costs for even a single transporta
tion infrastructure project is often highly time consum
ing or simply impossible.
This state of affairs explains why small-sample stud
ies dominate scholarship in this field of research. But de
spite the problems mentioned, after 4 years of data col
lection and refinement, we were able to establish a sample
of 258 transportation infrastructure projects with data
on both actual construction costs and estimated costs at
the time of decision to build. The project portfolio is
worth approximately US$90 billion (1995 prices). The
project types are bridges, tunnels, highways, freeways,
high-speed rail, urban rail, and conventional (interurban)
rail. The projects are located in 20 countries on 5 conti
nents, including both developed and developing nations.
The projects were completed between 1927 and 1998.
Older projects were included in the sample in order to
test whether the accuracy of estimated costs improved
over time. The construction costs of projects range from
US$1.5 million to US$8.5 billion (1995 prices),with the
smallest projects typically being stretches of roads in
larger road schemes, and the largest projects being rail
links, tunnels, and bridges. As far as we know, this is the
largest sample of projects with data on cost development
that has been established in this field of research.
In statistical analysis, data should be a sample from
a larger population, and the sample should represent the
population properly. These requirements are ideally sat
isfied by drawing the sample by randomized lot. Ran
domization ensures with high probability that factors
that cannot be controlled are equalized. A sample should
also be designed such that the representation of sub
groups corresponds to their occurrence and importance
in the population. In studies of human affairs, however,
where controlled laboratory experiments often cannot
be conducted, it is frequently impossible to meet these
ideal conditions. This is also the case for the current
study, and we therefore had to take a different approach
to sampling and statistical analysis.
We selected the projects for the sample on the basis
of data availability. All projects that we knew of for which
data on construction cost development were obtainable
were considered for inclusion in the sample. Cost devel
opment is defined as the difference between actual and
estimated costs in percentage of estimated costs, with all
costs measured in fixed prices. Actual costs are defined
as real, accounted costs determined at the time of com
pleting a project. Estimated costs are defined as bud
geted, or forecasted, costs at the time of decision to build.
Even if the project planning process varies with project
type. country, and time, it is typically possible to locate
for any given project a specific point in the process that
can be identified as the time of decision to build. Usually
a cost estimate was available for this point in time. If not,
the closest available estimate was used, typically a later
estimate resulting in a conservative bias in our measure
ment of cost development. Cost data were collected from
a variety of sources, including annual project accounts,
questionnaires, interviews, and other studies.
APA Journal • Summer 2002 • Vol. 68, No. 3
293
BENT FLYVBJERG, METE SKAMRIS HOLM, AND SØREN BUHL
Data on cost development were available for 343
projects. We then rejected 85 projects because of insuffi
cient data quality. For instance, for some projects we
could not obtain a clear answer regarding what was in
cluded in costs, or whether cost data were given in cur
rent or fixed prices, or which price level (year) had been
used in estimating and discounting costs. More specifi
cally, of those 85 projects, we rejected 27 because we
could not establish whether or not cost data were valid
and reliable. We rejected 12 projects because they had
been completed before 1915 and no reliable indices were
available for discounting costs to the present. Finally, we
excluded 46 projects because cost development for them
turned out to have been calculated before construction
was completed and operations begun: therefore, the ac
tual final costs for these projects may be different from
the cost estimates used to calculate cost development,
and no information was available on actual final costs. In
addition to the 85 rejected projects mentioned here, we
also rejected a number of projects to avoid double count
ing of projects. This typically involved projects from
other studies that appeared in more than one study or
where we had a strong suspicion that this might be the
case. In sum, all projects for which data were considered
valid and reliable were included in the sample. This cov
ers both projects for which we ourselves collected the
data and projects for which other researchers in other
studies did the data collection (Fouracre et al., 1990:
Hall, 1980: Leavitt et al., 1993; Lewis, 1986; Merewitz,
1973a; National Audit Office, Department of Transport,
1985, 1992: National Audit Office, Department of
Transport, Scottish Development Department, & Welsh
Office, 1988; Pickrell, 1990; Riksrevisionsverket, 1994;
Vejdirektoratet, 1995: Walmsley & Pickett, 1992). Cost
data were made comparable across projects by discount
ing prices to the 1995 level and calculating them in
Euros, using the appropriate geographical, sectoral, and
historical indices for discounting and the appropriate
exchange rates for conversion between currencies.
Our own data collection concentrated on large Eu
ropean projects because too few data existed for this type
of project to allow comparative studies. For instance, for
projects with actual construction costs larger than 500
million Euros (1995 prices: EUR1=US$ 1.29 in 1995), we
were initially able to identify from other studies only two
European projects for which data were available on both
actual and estimated costs. If we lowered the project size
and looked at projects larger than 100 million Euros, we
were able to identify such data for eight European pro
jects. We saw the lack of reliable cost data for European
projects as particularly problematic since the Commis
sion of the European Union hadjust launched its policy
for establishing the so-called trans-European transport
294 APA Journal • Summer 2002 • Vol. 68, No. 3
networks, which would involve the construction of a
large number of major transportation infrastructure
projects across Europe at an initial cost of 220 billion
Euros (Commission of the European UnIon, 1993, p.
75). As regards costs, we concluded that the knowledge
base for the Commission’s policy was less than well de
veloped, and we hoped to help remedy this situation
through our data collection. Our efforts on this point
proved successful. We collected primary data on cost for
37 projects in Denmark, France, Germany, Sweden, and
the U.K. and were thus able to greatly increase the num
ber of large European projects with reliable data for both
actual and estimated costs, allowing for the first time a
comparative study for this type of project in which sta
tistical methods could be applied.
As for any sample, a key question is whether the sam
ple is representative of the population. Here the ques
tion is whether the projects included in the sample are
representative of the population of transportation in
frastructure projects. Since the criterion for sampling
was data availability, this question translates into one of
whether projects with available data are representative.
There are four reasons why this is probably not the case.
First, it may be speculated that projects that are man
aged well with respect to data availability may also be
managed well in other respects, resulting in better than
average (i.e., nonrepresentative) performance for such
projects. Second, it has been argued that the very exis
tence of data that make the evaluation of performance
possible may contribute to improved performance when
such data are used by project management to monitor
projects (World Bank, 1994, p. 17). Again, such projects
would not be representative of the project population.
Third, we might speculate that managers of projects
with a particularly bad track record regarding cost esca
lation have an interest in not making cost data available,
which would then result in underrepresentation of such
projects in the sample. Conversely, managers of projects
with a good track record for costs might be interested in
making this public, resulting in overrepresentation of
these projects. Fourth, and finally, even where managers
have made cost data available, they may have chosen to
give out data that present their projects in as favorable a
light as possible. Often there are several estimates of
costs to choose from and several calculations of actual
costs for a given project at a given time. If researchers col
lect data by means of survey questionnaires, as is often
the case, there might be a temptation for managers to
choose the combination of actual and estimated costs
that suits them best, possibly a combination that makes
their projects look good.
The available data do not allow an exact, empirical
assessment of the magnitude of the problem of misrep
UNDERESTIMATING COSTS IN PUBLIC WORKS PROJECTS
resentation. But the few data that exist that shed light on
this problem support the thesis that data are biased.
When we compared data from the Swedish Auditor Gen
eral for a subsample of road projects, for which the prob
lems of misrepresentation did not seem to be an issue,
with data for all road projects in our sample, we found
that cost escalation in the Swedish subsample is signifi
cantly higher than for all projects (Holm, 1999, pp. 11—
15). We conclude, for the reasons given above, that most
likely the sample is biased and the bias is conservative. In
other words, the difference between actual and estimated
costs derived from the sample is likely to be lower than
the difference in the project population. This should be
kept in mind when interpreting the results from statisti
cal analyses of the sample. The sample is not perfect by
any means. Still it is the best obtainable sample given the
current state of the art in this field of research.
In the statistical analyses, percentage cost develop
ment in the sample is considered normally distributed
unless otherwise stated. Residual plots, not shown here,
indicate that normal distribution might not be com
pletely satisfied, the distributions being somewhat
skewed with larger upper tails. However, transforma
tions (e.g., the logarithmic one) do not improve this sig
nificantly. For simplicity, therefore, no transformation
has been made, unless otherwise stated.
The subdivisions of the sample implemented as part
of analyses entail methodological problems of their own.
Thus the representation of observations in different
combinations of subgroups is quite skewed for the data
considered. The analysis would be improved consider
ably if the representation were more even. Partial and
complete confounding occur; that is, if a combination
of two or more effects is significant, it is sometimes dif
ficult to decide whether one, the other, or both cause the
difference. For interactions, often not all the combina
tions are represented, or the representations can be quite
scarce. We have adapted our interpretations of the data
to these limitations, needless to say. If better data could
be gathered, sharper conclusions could be made.
The statistical models used are linear normal models
(i.e., analysis of variance and regression analysis with the
appropriate F-tests and t-tests). The tests of hypotheses
concerning mean values are known to be robust to devi
ations from normality. Also, chi-square tests for inde
pendence have been used for count data. For each test,
the p-value has been reported. This value is a measure
for rareness if identity of groups is assumed. Tradition
ally, a p-value less than 0.01 is considered highly signifi
cant and less than 0.05 significant, whereas a larger p
value means that the deviation could be due to chance.
APA Journal • Summer 2002 • Vol. 68, No. 3 295