Abstracts

Transcription

Abstracts
Pre-Conference Tutorials — Room 2000
Pre-Conference Tutorials — Room 2009
8:00 a.m.
8:00 a.m.
Optimizing Big Data Programs
Model Selection with SAS/STAT® Software
This seminar will review traditional SAS techniques for optimizing programs
that require table lookups. We will do this by focusing on in-memory
techniques. The techniques discussed will include: Arrays (one-dimensional
and multi-dimensional). Hash objects. Formats. We will look at indexing as a
means of “super-charging” your lookups and sampling for the purpose of
exploring your data. We will also investigate several newer features in SAS
as a way to optimize your programs in a BI environment, including grid
computing and the use of PROC SCAPROC to help “gridify” your programs.
This is an extra fee event. See registration.
When you are faced with a predictive modeling problem that has many
possible predictor effects – dozens, hundreds or even thousands – a natural
question is, "What subset of the effects provides the best model for the
data?" This tutorial explores how you can use model selection methods in
SAS/STAT software to address this question. Although the model selection
question seems reasonable, trying to answer it for real data can create
problematic pitfalls, leading some experts to stridently denounce model
selection. This workshop focuses on the GLMSELECT procedure and shows
how it can be used to address and mitigate the intrinsic difficulties of model
selection. You will learn how to:
Christine Riddiough, SAS
Pre-Conference Tutorials — Room 2007
8:00 a.m.
Modeling Categorical Response Data
Maura Stokes, SAS
Logistic regression, generally used to model dichotomous response data, is
one of the basic tools for a statistician. But what do you do when maximum
likelihood estimation fails or your sample sizes are questionable? And
wWhat happens when you have more than two response levels? And how
do you handle counts? This tutorial briefly reviews logistic regression for
dichotomous responses, and then illustrates alternative strategies for the
dichotomous case and additional strategies such as the proportional odds
model, generalized logit model, conditional logistic regression, and Poisson
regression. . The presentation is based on the third edition of the book
Categorical Data Analysis Using the SAS System by Stokes, Davis and Koch
(2012). An existing working knowledge of logistic regression is required for
this tutorial to be fully beneficial. This is an extra fee event. See registration.
Pre-Conference Tutorials — Room 2008
8:00 a.m.
A-to-Z Analysis and Validation Using PROC Tabulate
Sunil Gupta, Gupta Programming
Generate most any combination of summary table layout! Do you know up
to seven different table layouts, five ways to control order, three ways to
include missing data, or four ways to include subtotals? After this course,
you can expect to take advantage of PROC Tabulate’s powerful feature to
group and analyze data in most any summary table layout. This unique
course explores core syntax options for up to 26 key summary table
structures, including mixing both continuous and categorical data with
ODS. By applying the simple SAS examples provided throughout the
course, you too can master PROC Tabulate in your daily programming
environment. Each student receives the companion SAS e-guide, which is a
great reference tool for searching, cutting and pasting concise model SAS
examples. This is an extra fee event. See registration
Funda Gunes, SAS
• Use extensive model selection diagnostics including graphics to detect
problems.
• Use validation data to detect and prevent both under- and over-fitting.
• Use modern-penalty based methods, including LASSO and adaptive
LASSO, as alternatives to traditional methods such as stepwise selection.
• Use bootstrap-based model averaging to reduce selection bias and
improve predictive performance. This tutorial requires an understanding
of basic regression techniques. This is an extra fee event. See registration
Pre-Conference Tutorials — Room 2010
8:00 a.m.
How to Become a Top SAS Programmer
Michael Raithel, Westat
This groundbreaking seminar, based on the new SAS Press book with the
same title, provides clear-cut strategies for becoming a top SAS
programmer. Whether you are a student or a statistician, a programmer or a
business analyst, this seminar shows how you can streamline and revitalize
your career to become the top SAS professional in your organization.
Instructor Michael A. Raithel will reveal how to unleash the SAS expert
within you by learning how to take advantage of a wide variety of proven
job strategies that use SAS to help you to maximize your knowledge, skills,
accomplishments, recognition and pay. Featuring lectures, discussions and
paper-and-pen exercises, the seminar details key SAS programming
fundamentals to master, strategies for becoming the go-to SAS person in
your organization, ways to become involved and recognized in the greater
SAS community, and how to exploit the best sources of SAS information.
You will leave the seminar with a written set of objectives and the
knowledge to implement them. Armed with the information from this
course, your set of goals and objectives, and your own ambition, there are
no limits to how far you can go with your SAS programming career. This is
an extra fee event. See registration.
www.sasglobalforum.org/2013
1
Pre-Conference Tutorials — Room 2012
Pre-Conference Tutorials — Room 2009
8:00 a.m.
10:30 a.m.
SAS Enterprise BI – Tables, Maps, and Cubes:
Understanding the Differences
Introduction to the MCMC Procedure in SAS/STAT
Software
Eric Rossland, SAS
Fang Chen, SAS
You have several choices of data sources when analyzing data and creating
reports with the SAS BI applications. Whether you’re new to SAS BI or have
been using it since the beginning, this seminar will help you to understand
the different types of data sources. Tables, OLAP cubes, and information
maps can all provide data to the SAS Add-In for Microsoft Office, SAS
Enterprise Guide, and SAS Web Report Studio. Each of these data sources
has unique characteristics that can enhance the analysis and reporting. Join
us as we investigate each data source and how it can be used in the various
SAS applications. Intended Audience: Anyone who has SAS Enterprise BI
Server or wants to learn more about the different data sources as well as
the analytic and reporting capabilities they provide. This is an extra fee
event. See registration.
The MCMC procedure is a general-purpose Markov chain Monte Carlo
simulation tool designed to fit Bayesian models. It uses a variety of
sampling algorithms to generate samples from the targeted posterior
distributions. This workshop will review the methods available with PROC
MCMC and demonstrate its use with a series of real-world applications.
Examples will include fitting a variety of parametric models: generalized
linear models, linear and nonlinear models, hierarchical models, zeroinflated Poisson models, and missing data problems. Additional Bayesian
topics such as sensitivity analysis, inference of functions of parameters, and
power priors will be discussed, and applications will be demonstrated with
the MCMC procedure. This workshop is intended for statisticians who are
interested in Bayesian computation. Attendees should have a basic
understanding of Bayesian methods and experience using the SAS
language. This tutorial is based on SAS/STAT 12.1. This is an extra fee event.
See registration
Pre-Conference Tutorials — Room 2007
10:30 a.m.
Creating Statistical Graphics in SAS
Warren Kuhfeld, SAS
Effective graphics are indispensable in modern statistical analysis. SAS
provides statistical graphics through ODS Graphics, functionality that is
used by statistical procedures to create statistical graphics as automatically
as they create tables. ODS Graphics is also used by a family of Base SAS
procedures designed for graphical exploration of data. This tutorial is
intended for statistical users and covers the use of ODS Graphics from start
to finish in statistical analysis. You will learn how to:
• Request graphs created by statistical procedures.
• Use the SGPLOT, SGPANEL, SGSCATTER and SGRENDER procedures in
SAS/GRAPH® to create customized graphs.
• Access and manage your graphs for inclusion in Web pages, papers and
presentations.
• Modify graph styles (colors, fonts and general appearance).
• Make immediate changes to your graphs using a point-and-click editor.
Pre-Conference Tutorials — Room 2000
12:30 p.m.
Beyond the Basics: Advanced Macro Tips and
Techniques
Jim Simon, SAS
This seminar will investigate ways to automate common programming
tasks through the use of advanced macro techniques. We will investigate
different techniques for generating repetitive data-driven macro calls,
including the use of the EXECUTE routine. We will look at how to use
various I/O functions to access SAS data sets and create your own macro
functions. We will also examine expediting the process of importing
external files such as CSV and Excel files. Along the way, we will learn
techniques for validating user input. This is an extra fee event. See
registration.
• Make permanent changes to your graphs with template changes.
Pre-Conference Tutorials — Room 2008
• Specify other options related to ODS Graphics. This is an extra fee event.
See registration
12:30 p.m.
What Will DS2 Do for You?
Mark Jordan, SAS
DS2 is an exciting, powerful new programming language available in SAS
9.4. It enables users to explicitly control threading to leverage multiple
processors when performing data manipulation and data modeling. It also
improves extensibility and data abstraction in your code through the
implementation of packages and methods. Paired with the SAS Embedded
Process, DS2 enables you to perform processing similar to SAS in
completely new places, such as in-database processing in relational
databases, the SAS High-Performance Analytics grid and the DataFlux®
Federation Server. In this seminar, you will learn the basics of writing and
2
www.sasglobalforum.org/2013
executing DS2 code. Discover how you can capitalize on the power of the
new DS2 programming language in your own work. This is an extra fee
event. See registration.
columns, some of which can cause unexpected results for the unwary user.
And after all, PROC SQL is part of Base SAS; so, though you may need to
learn a few new keywords to become an SQL wizard, no special license is
required! This is an extra fee event. See registration.
Pre-Conference Tutorials — Room 2010
12:30 p.m.
Output Delivery System: The Basics and Beyond
Kirk Paul Lafler, Software Intelligence Corporation
This course explores the various techniques associated with output
formatting and delivery using the Output Delivery System (ODS). Numerous
examples will be presented to command mastery of ODS capabilities while
providing a better understanding of ODS statements and options to deliver
output any way that is needed. Topics include:
• SAS-supplied formatting statements and options.
• Selecting output objects with Selection or Exclusion Lists.
• Formatting Output as RTF, PDF, Microsoft Excel, and HTML.
• Using the Escape character to enhance output formats.
• Exploring ODS statements and options.
• Implementing scrollable tables in HTML output with static column
headers.
• Enabling/disabling borders.
• Generating HTML hyperlinks in RTF output.
• Adding images to RTF output.
• Removing gridlines and shading in RTF output.
• Creating a printable table of contents in PDF output.
• Sending output to Microsoft Office.
• Constructing drill-down applications with the DATA step, ODS and SAS/
GRAPH software.
• Creating thumbnail charts.
• Techniques for creating user-defined ODS styles.
• An introduction to the customization of output with the TEMPLATE
Procedure. This is an extra fee event. See registration
Pre-Conference Tutorials — Room 2012
12:30 p.m.
Demystifying PROC SQL
Christianna Williams, Independent Consultant
Subqueries, inline views, outer joins, Cartesian products, HAVING
expressions, Set operators, INTO clauses – even the terminology of SQL can
be rather daunting for SAS programmers who use DATA step for data
manipulation. Not to mention the profusion of commas and complete
dearth of semicolons found in a PROC SQL step! Nonetheless, even the
most die-hard DATA step programmers must grudgingly acknowledge that
there are some tasks – such as the many-to-many merge or the “not-quiteequi-join” – that would require Herculean effort to accomplish with DATA
steps. However, these tasks can be achieved amazingly concisely, even
elegantly, using PROC SQL. This seminar will present a series of increasingly
complex examples to illustrate the function of each of PROC SQL’s clauses,
with particular focus on summarization/aggregation and a variety of joins.
Additionally, the examples will illuminate how SQL “thinks” about rows and
www.sasglobalforum.org/2013
3
4
www.sasglobalforum.org/2013
Applications Development — Room 2014
4:00 p.m.
2:00 p.m.
Fueling the Future of an Energy Company
Stijn Vanderlooy, EDF Luminus
Floating on Cloud 9.3: Leveraging the Cloud with SAS®
and Google Drive
William Roehl, MarketShare
Paper 001-2013
Our organization has been utilizing Google Drive (previously Google Docs)
to keep project documentation centrally located for ease of access by any
user on any platform. Up to this point, SAS® developers had to manually
import or export data sets to or from flat files or Microsoft Excel in order to
update data stored in the cloud. This paper provides a powerful macro
toolkit that facilitates direct access to Google Spreadsheets through its
published API, allowing uploading, downloading, deletion, and live data
manipulation of cloud-based data. This provides an opportunity for a
significant reduction in the amount of manual work and time required for
SAS developers to perform these basic functions. Code was developed with
SAS 9.3 and HTMLTidy (25MAR2009) running under Microsoft Windows 7.
2:30 p.m.
Using PROC FCMP in SAS® System Development: Real
Examples
Xiyun (Cheryl) Wang, Statistics Canada
Yves Deguire, Statistics Canada
Paper 505-2013
This paper discusses the use of the FCMP procedure, focusing mainly on
two aspects. One aspect is that a lot of complex mathematical functions to
be used in our systems are difficult to implement using SAS® macros. I have
encapsulated these complex functions into PROC FCMP functions and used
them seamlessly in PROC OPTMODEL. Another aspect is the need to apply
generic message handling across different SAS components, such as the
DATA step, SAS macros, and PROC steps. PROC FCMP again becomes a
natural fit for this purpose that can be easily invoked by any SAS program
blocks.
3:00 p.m.
Create Your Own Client Apps Using SAS® Integration
Technologies
Chris Hemedinger, SAS
Paper 003-2013
SAS® Integration Technologies allows any custom client application to
interact with SAS services. SAS® Enterprise Guide® and SAS® Add-In for
Microsoft Office are noteworthy examples of what can be done, but your
own applications don't have to be that ambitious. This paper explains how
to use SAS Integration Technologies components to accomplish focused
tasks, such as run a SAS® program on a remote server, read a SAS data set,
run a stored process, and transfer files between the client machine and the
SAS server. Working examples in Microsoft .NET (including C# and Visual
Basic .NET) as well as Windows PowerShell are also provided.
Paper 004-2013
EDF Luminus is a producer and supplier of energy in Belgium. The company
is active in several markets for trading commodities (including electricity,
natural gas, and oil-related products). The market data modeling team is
responsible for all information related to these markets. Besides a daily
collection and verification of all the published prices, the team is
responsible for a diverse set of transformations and manipulations of these
data. Examples are numerous and include volatility estimation and price
forecasts. In this paper we present a successful application of a SAS® tool
developed in-house that is used by the market data modeling team to
support its core tasks. The tool runs autonomously three times a day.
4:30 p.m.
Macro Quoting to the Rescue: Passing Special Characters
Art Carpenter, CA Occidental Consultants
Mary Rosenbloom, Edwards Lifesciences, LLC
Paper 005-2013
We know that we should always try to avoid storing special characters in
macro variables. We know that there are just too many ways that special
characters can cause problems when the macro variable is resolved.
Sometimes, however, we just do not have a choice. Sometimes the
characters must be stored in the macro variable whether we like it or not.
And when they appear we need to know how to deal with them. We need
to know which macro quoting functions will solve the problem, and even
more importantly why we need to use them. This paper takes a quick look
at the problems associated with the resolution and use of macro variables
that contain special characters such as commas, quotes, ampersands, and
parentheses.
5:00 p.m.
Line-Sampling Macro for Multistage Sampling
Charley Jiang, University of Michigan
James Lepkowski, University of Michigan
Richard Valliant, University of Michigan
James Wagner, University of Michigan
Paper 007-2013
In the SAS® world, one tool, PROC SURVEYSELECT, is widely used for
probability sample selection. However, the procedures implemented in
PROC SURVEYSELECT are basic selection tools that must be assembled into
larger systems for complex probability samples, particularly multistage
samples. This paper describes the development and operation of a set of
sampling macros built around PROC SURVEYSELECT for sampling the
ultimate stage units in a multistage sample. Examples of several situations
where the macro can be most beneficial are also given.
5:30 p.m.
SAS® Analytics Optimized with Intel Technologies
Mark Pallone, Intel Corporation
(Invited) Paper 539-2013
Several server platform configurations and technologies have a direct
impact on performance and scalability that are critical for SAS® workloads.
Intel Cache Acceleration Solution (iCAS) reduces storage latency and
transparently accelerates Applications, Servers, and Virtual Machines. iCAS
www.sasglobalforum.org/2013
5
was used to measure IO performance improvements for SAS mixed
analytics workload. Resulting IO performance metrics will be shared.
Performance testing was conducted on 3rd Generation Intel Core processor
family server platform that is expected to ship in Q3 2013. Comparison of
performance metrics between 2nd and 3rd Generation Intel Core processor
families based on mixed analytics workload will be reviewed. In addition,
some preliminary performance metrics on 4th generation Intel iCore
processor will be shared as well.
Beyond the Basics — Room 2016
10:30 a.m.
Using the SAS® Data Step to Generate HTML or TextBased Mark-Up
Matt Karafa, The Cleveland Clinic
(Invited) Paper 020-2013
The author presents macros which produce reports direct to MS Wordcompliant HTML, thus demonstrating an alternative method to create MS
Word documents from SAS®. The first step is to create a mock-up of the
table in an external mark-up editor, then use SAS to produce the text that
creates the file, interspersing the required data between the mark-up tags.
These macros demonstrate a way to increase the control and flexibility over
what is available via the traditional ODS RTF or HTML mechanism. Further,
via this method, any text-based mark-up language (HTML, RTF, LaTeX, etc.)
can be produced with a minimal effort.
11:30 a.m.
The Hash-of-Hashes as a "Russian Doll" Structure: An
Example with XML Creation
Joseph Hinson, MERCK
Paper 021-2013
SAS®9 hash objects have inspired novel programming techniques. The
recent discovery that hash tables can contain even other hash objects:
“hash of hashes” opens the door to their application to hierarchical data
processing. Because hierarchies, like Russian dolls, can be considered
"containers within containers.” Thus, nested hash objects could model XML,
a hierarchical data structure increasingly finding its way into clinical trial
data. Clinical programmers now have to deal with hierarchical as well as
tabular and relational data sets. SAS® now provides tools like the XML
libname engine and XML mapper. This paper aims to show, using a
simplified CDISC LAB model, that the hash object could well be another
tool for creating XML.
12:00 p.m.
Optimize Your Delete
Brad Richardson, SAS
Paper 022-2013
Have you deleted a data set or two from a library that contains thousands
of members, using PROC DATASETS? If so, you probably have witnessed
some wait time. To maximize performance, we have reinstated PROC
DELETE as a supported procedure. One of the main differences between
PROC DELETE and PROC DATASETS DELETE is that PROC DELETE does not
need an in-memory directory to delete a data set. So what does this mean
exactly? This presentation explains all. For PROC DATASETS DELETE fans,
there are optimizations, as well. Come to this presentation to learn more
from the developer.
6
www.sasglobalforum.org/2013
2:00 p.m.
Using Mail Functionality in SAS®
Erik Tilanus, Synchrona
(Invited) Paper 023-2013
The DATA step can write information to virtually any destination. You are
probably familiar with writing to SAS data sets and external files. But also
email can function as a destination. The paper will discuss how to configure
the email features in the system options and share practical hints how to
use them. Then we will proceed with sending a simple email from the DATA
step, with or without attachments. Then we will use extensions to the
standard PUT statement to support the email facility to send personalized
mass-mailings. Finally, we will show how to send procedure output that has
been created using ODS.
3:00 p.m.
The Ins and Outs of Web-Based Data with SAS®
Bill McNeill, SAS
Paper 024-2013
Do you have data on the Web that you want to integrate with SAS®? This
paper explains how you can obtain Web data, process it and export it back
out to the Web. Examples will use existing features, such as the SOAP and
XSL procedures, the XML mapper application, and XMLV2 LIBNAME engine,
along with two new features: the XMLV2 LIBNAME engine AUTOMAP
option and the JSON procedure. The AUTOMAP option allows for creation
of default XML Mapper files within SAS. The JSON procedure exports SAS
data sets in JSON format to an external file. And if you need to write freeform JSON output, forget the SAS PUT statements; the JSON procedure
supports free-form JSON output as well.
4:00 p.m.
Internationalization 101: Give Some International Flavor
to Your SAS® Applications
Mickael Bouedo, SAS
Steve Beatrous, SAS
Paper 025-2013
Do you have SAS® users worldwide? Do you want your SAS application to
be useable in many languages? SAS® 9.4 internationalization features will
get you there efficiently. If you want to adapt your SAS application for
different cultures, SAS internationalization is the step which generalizes
your product to be language-independent. Internationalization features in
SAS include the ENCODING, LOCALE, and TIMEZONE options; the SASMSG
function; the NL formats; and many more features that help you to write
code once so it can run in different cultural environments with low
maintenance. This paper describes how to successfully internationalize
your SAS programs and make them ready for the world.
4:30 p.m.
ISO 101: A SAS® Guide to International Dating
Peter Eberhardt, Fernwood Consulting Group Inc
Xiaojin Qin, Covance Pharmaceutical Research and
Development CO., Ltd.
(Invited) Paper 026-2013
For most new SAS® programmers, SAS dates can be confusing. Once some
of this confusion is cleared, the programmer might then come across the
ISO date formats in SAS, and another level of confusion sets in. This paper
reviews SAS date, SAS datetime, and SAS time variables and some of the
ways they can be managed. It then turns to the SAS ISO date formats and
shows how to make your dates international.
5:30 p.m.
Census Retires PROC COMPUTAB
Christopher Boniface, U.S. Census Bureau
Nora Szeto, U.S. Census Bureau
Hung Pham, U.S. Census Bureau
Paper 027-2013
PROC COMPUTAB is used to generate tabular reports in a spreadsheet-like
format. PROC COMPUTAB has been around a long time. It has served us
well at Census, but it is time to replace it with reporting procedures that are
more modern. This paper shows you how to create hundreds of Excel tables
using ODS TAGSETS EXCELXP. We discuss how we converted PROC
COMPUTAB to both PROC TABULATE and PROC REPORT to create complex
Census tables. Moreover, how we use PROC TABULATE as the computing
engine to handle overlapping format ranges and PROC REPORT as the
reporting tool to create polished Excel tables. We reveal how to control the
appearance of the Excel tables including column widths, row heights, and
formats.
allocation decisions become significant. This paper describes how we are
using SAS® Enterprise Miner™ to develop a model to score university
students based on their probability of enrollment and retention early in the
enrollment funnel so that staff and administrators can work to recruit
students that not only have an average or better chance of enrolling but
also succeeding once they enroll. Incorporating these results into SAS® EBI
will allow us to deliver easy-to-understand results to university personnel.
12:00 p.m.
Using SAS® BI for Integrated Bank Reporting
James Beaver, Farm Bureau Bank
Paper 045-2013
This paper shows how Base SAS®, SAS® Enterprise Guide®, SAS/ETS®, and
SAS® BI are used to provide a comprehensive view of bank performance.
Data is extracted from the G/L, loan, deposit, and application systems, realtime data is accessed to provide up-to-the-minute results on loan activity,
and system reports are read in to provide additional information. PROC
COMPUTAB is used to create financial statements, OLAP cubes are used to
provide reports on bank balance sheet components and budget
comparisons on non-interest income and expense items by department,
and dashboards are used to provide real-time reports on loan originations.
The reports are presented using SAS BI through the SAS data portal to
provide real-time, trend, and historical reports on the bank’s performance.
Business Intelligence Applications — Room 2009
2:00 p.m.
10:30 a.m.
SAS® High-Performance Analytics: Big Data Brought to
Life on the EMC Greenplum Data Computing Appliance
SAS® Business Intelligence Development Roundtable:
SAS Business Intelligence Solutions Portfolio and Future
Focus
Greg Hodges, SAS
Stuart Nisbet, SAS
Don Chapman, SAS
James Holman, SAS
Oita Coleman, SAS
Tammi Kay George, SAS
Paper 063-2013
Join SAS executives from Product Management and Business Intelligence
Research & Development for an interactive panel discussion on the current
BI and reporting solutions portfolio including SAS Enterprise BI Server, SAS
Enterprise Guide, Mobile BI and SAS Visual Analytics. These experts will
answer questions such as how to decide which product to use when, offer
deployment and implementation best practices and also provide guidance
on strategies to add Visual Analytics to your mix of SAS solutions. High level
roadmaps will be shared by product management that cover the product
portfolio and there will be dedicated time for Q&A to make sure your
questions get answered!
11:30 a.m.
A Data-Driven Analytic Strategy for Increasing Yield and
Retention at Western Kentucky University Using SAS
Enterprise BI and SAS® Enterprise Miner™
Matt Bogard, Western Kentucky University
Paper 044-2013
Paul Cegielski, Greenplum
(Invited) Paper 064-2013
This presentation will describe the proof-of-concept project to apply highperformance analytics (HPA) to call center and other data in an effort to
quickly identify and act on customer service opportunities. Discussion will
include functionality and performance metrics of SAS® High-Performance
Analytics procedures, the new SAS® DS2 language, the fast-loading
capability of the Greenplum DCA, and the ability to deploy models built on
the DCA to other databases. Since some of the most valuable data is
unstructured, such as the free-form text notes entered by call center staff,
the presentation will describe how SAS® Text Miner is used in conjunction
with the HPA DCA to include unstructured data in analyses and modeling.
3:00 p.m.
SAS® Stored Processes Are Goin’ Mobile!: Creating and
Delivering Mobile-Enabled Versions of Stored Process
Reports
Michael Drutar, SAS
Paper 046-2013
SAS® BI programmers have been clamoring for a way to quickly create
mobile-enabled versions of existing SAS BI content. This can be difficult
because there may be multiple BI reports that need to be converted.
Fortunately, PROC STP (new in SAS 9.3) is a solution that can take a system’s
existing SAS Stored Process reports and create mobile-enabled versions of
them. This paper shows how PROC STP can capture the ODS HTML output
from a stored process, create an HTML file from it and email the file to any
mobile device (iPhone®, Android, etc.). Using this method, any existing SAS
Stored Process report’s output can be easily mobile-enabled.
As many universities face the constraints of declining enrollment
demographics, pressure from state governments for increased student
success, as well as declining revenues, the costs of utilizing anecdotal
evidence and intuition based on “gut” feelings to make time and resource
www.sasglobalforum.org/2013
7
3:30 p.m.
Linking Strategy Data in BI Applications
these features to enhance basic OLAP cubes by using member properties,
defining dynamic measures and dimensions using the MDX language, and
improving performance for high data volumes.
Paper 050-2013
Data Management — Room 2001
Bharat Trivedi, SAS
Christiana Lycan, SAS
SAS® Enterprise BI is a key part of every strategy management
implementation, but it is not always easy to link strategy data with the
correct views of information in SAS Enterprise BI. And even when the
linking is straightforward, users may not be authorized to see all of the
information. As a result, several views of the same reports must be created
for linking. Users need an easier way to link and display the correct view of
strategy management content in SAS Enterprise BI. This paper reveals how
to link information created in SAS Enterprise BI that is context-aware and
secure.
4:00 p.m.
Self-Service Data Management: Visual Data Builder
Malcolm Alexander, SAS
Sam Atassi, SAS
Paper 051-2013
Successful data preparation is the key to extracting meaningful knowledge
from data. SAS® Visual Data Builder allows you to access data from
enterprise sources and transform it for use in business intelligence, data
visualization and data mining tasks. This paper discusses self-service data
management techniques available using SAS Visual Data Builder, as well as
the unique features enabling it to load data into SAS® LASR(TM)
Analytic(TM) Server.
Business Intelligence Applications — Room 3016
4:30 p.m.
Stop your 'Wine”-ing: Use a Stored Process!
Tricia Aanderud, And Data Inc
Angela Hall, SAS
(Invited) Paper 043-2013
One of the major benefits of using SAS® Stored Processes is extendibility.
SAS® stored processes are one of the most customizable products; there are
several advantages, such as the ability to set up reports that can run in
various locations, enhance out-of-the box functionality with custom
widgets, and leverage all of the stored process server options. In this
discussion, you will learn advanced tips and tricks for using stored
processes within SAS BI clients.
5:30 p.m.
Escape from Big Data Restrictions by Leveraging
Advanced OLAP Cube Techniques
Stephen Overton, Overton Technologies LLC
Paper 047-2013
In today’s fast-growing field of business analytics, there are many tools and
methods for summarizing and analyzing big data. This paper focuses
specifically on OLAP technology and features natively available in OLAP
cubes that enable organizations to deploy robust business intelligence
reporting when high volumes of data exist. This paper discusses how to use
8
www.sasglobalforum.org/2013
2:00 p.m.
What's New in SAS® Data Management
Malcolm Alexander, SAS
Nancy Rausch, SAS
Paper 070-2013
The latest releases of SAS® Data Integration Studio and SAS® Data
Management provide an integrated environment for managing and
transforming your data to meet new and increasingly complex data
management challenges. The enhancements help develop efficient
processes that can clean, standardize, transform, master, and manage your
data. Latest features include capabilities for building complex job
processes; new web-based development and job-monitoring
environments; enhanced ELT transformation capabilities; big data
transformation capabilities for Hadoop; integration with the SAS® LASR™
platform; enhanced features for lineage tracing and impact analysis; and
new features for master data and metadata management. This paper
provides an overview of the latest features of the products and includes use
cases and examples of the product capabilities.
3:00 p.m.
Bigger Data Analytics: Using SAS® on Aster Data and
Hadoop
John Cunningham, Teradata
Paper 071-2013
With the increased popularity of new Big Data clustered processing
platforms, SAS® Analytics now has the opportunity to solve newer, bigger
problems than ever before. Paper will focus on the evolution of Big Data
analytics, the new data sources and types, new technologies involved, to
achieve end to end analytic processing with SAS. Will specifically
demonstrate the use of new Big Data technologies, SAS Analytics with SAS/
ACCESS® for Aster, Aster SQL-MR, SQL-H to integrate end to end Big Data
analytics on the Aster Discovery Platform, even from raw data files stored
on Hadoop clusters.
3:30 p.m.
SAS-Oracle Options and Efficiency: What You Don't
Know Can Hurt You
John Bentley, Wells Fargo Bank
(Invited) Paper 072-2013
SAS/Access engines allow SAS to read, write, and alter almost any relational
database. Using the engine right out of the box works OK, but there are a
host of options that if properly used can improve performance, sometimes
greatly. In some cases though an incorrect option value will degrade
performance. This paper will review cross-database SAS/Access engine
options that can impact performance. Examples and test cases using an
Oracle database will be provided. All levels of SAS programmers, Enterprise
Guide users, and non-Oracle database users will find the paper useful.
4:30 p.m.
Data Mining and Text Analytics — Room 2004
SAS® Data Management Techniques: Cleaning and
Transforming Data for Delivery of Analytic Data Sets
3:00 p.m.
Chris Schacherer, Clinical Data Management Systems, LLC
Paper 540-2013
The analytic capabilities of SAS® software are unparalleled. Similarly, the
ability of the Output Delivery System to produce an endless array of
creative, high-quality reporting solutions is the envy of many analysts using
competing tools. Beneath the glitz and glitter is the dirty work of cleaning,
managing, and transforming raw source data and reliably delivering
analytic data sets that accurately reflect the processes being analyzed.
Although a basic understanding of DATA step processing and PROC SQL is
assumed, the present work provides examples of both basic data
management techniques for profiling data as well as transformation
techniques that are commonly utilized in the creation of analytic data
products. Examples of techniques for automating the generation and
delivery of production-quality, enterprise-level data sets are provided.
5:00 p.m.
How to Do a Successful MDM Project in SAP Using SAS®
MDM Advanced
Casper Pedersen, SAS Denmark
Paper 074-2013
How difficult would it be to embark on master data management (MDM)
projects at a large SAP organization? SAP organizations often come with
lots of opinionated people, ambitious project plans, and vast complexities
and politics. Don’t go with the Big Bang approach; instead, try a controlled
evolution. Come and see how one approach to the KNA1 (Customer Master)
table and MARA (Material Data) table was implemented.
Data Mining and Text Analytics — Room 3016
2:00 p.m.
Using Data Mining in Forecasting Problems
Timothy Rey, Dow Chemical Company
Chip Wells, SAS
Justin Kauhl, Tata Consulting
(Invited) Paper 085-2013
In today’s ever-changing economic environment there is ample
opportunity to leverage the numerous sources of time series data now
readily available to the savvy business decision maker. Time series data can
be used for business gain if the data is converted to information and then
knowledge. Data mining processes, methods, and technology oriented to
transactional-type data (data not having a time series framework) have
grown immensely in the last quarter century. There is significant value in
the interdisciplinary notion of data mining for forecasting when used to
solve time series problems. The presentation describes how to get the most
value out of the myriad of available time series data by utilizing data mining
techniques specifically oriented to data collected over time.
Time Is Precious, So Are Your Models: SAS® Provides
Solutions to Streamline Deployment
Jonathan Wexler, SAS
Wayne Thompson, SAS
Paper 086-2013
Organizations spend a significant amount of time, often too much,
operationalizing models. The more time you can spend on analytics, and
the less time on deployment headaches, the better chance you have to
address core business challenges. This paper shows you how SAS has
accelerated data mining model deployment throughout the analytical life
cycle, by providing key integration points across SAS® solutions. Whether
you built your models using SAS® High-Performance Analytics Server, SAS®
Enterprise Miner™, or SAS/STAT®, this paper shows you how to automate
the management, publishing, and scoring of models by using SAS® Data
Integration Studio, SAS® Model Manager, and SAS® Scoring Accelerator.
Immediate benefits include reduced data movement, increased
productivity of analytic and IT teams, and faster time to results.
4:00 p.m.
Bringing Churn Modeling Straight to the Source: SAS®
and Teradata In-Database Model Development
Karl Krycha, Teradata
Jonathan Wexler, SAS
Paper 087-2013
This paper takes a closer look at the opportunities of using the predictive
analytic power of SAS® together with the performance and scalability of
Teradata. Users will see how the SAS® Analytics Accelerator for Teradata
improves modeling speed from hours to seconds, allowing users to
produce more models faster. The SAS Analytics Accelerator eliminates data
movement by moving SAS analytic computations capabilities to the
Teradata database. The paper provides an overview of the available
procedures and uses a typical business application to illustrate the full endto-end process of analytic modeling within Teradata.
4:30 p.m.
Demand Forecasting Using a Growth Model and
Negative Binomial Regression Framework
Michelle Cheong, Singapore Management University
Cally Ong Yeru, Singapore Management University
Murphy Choy, Singapore Management University
Paper 088-2013
In this paper, we look at demand forecasting by using a growth model and
negative binomial regression framework. Using cumulative sales, we model
the sales data for different wristwatch brands and relate it to their sales and
growth characteristics. We apply clustering to determine the distinctive
characteristics of each individual cluster. Four different growth models are
applied to the clusters to find the most suitable growth model to be used.
After determining the appropriate growth model to be applied, we then
forecast the sales by applying the model to new products being launched
in the market and continue to monitor the model further.
www.sasglobalforum.org/2013
9
5:00 p.m.
11:30 a.m.
Using Classification and Regression Trees (CART) in SAS®
Enterprise Miner™ for Applications in Public Health
Double-Clicking a SAS® File: What Happens Next?
Paper 089-2013
Paper 115-2013
Leonard Gordon, University of Kentucky
Classification and regression trees (CART)—a non-parametric methodology
—were first introduced by Breiman and colleagues in 1984. In this paper
they are employed using SAS® Enterprise Miner™, and several examples are
given to demonstrate their use. CART are underused (especially in public
health), and they have the ability to divide populations into meaningful
subgroups that allow the identification of groups of interest and enhance
the provision of products and services accordingly. They can provide a
simple yet powerful analysis. This paper attempts to demonstrate their
value and thus encourage their increased use in data analysis.
Sandy Gibbs, SAS
Donna Bennett, SAS
5:30 p.m.
When you double-click a SAS® file on your desktop or in Windows Explorer,
which program launches—SAS or SAS® Enterprise Guide®? Does the SAS
program open in a preferred application? Both Microsoft and SAS have
introduced changes in how you change the default SAS file type
associations, starting with Microsoft Windows Vista and with SAS 9.2. If you
have installed SAS 9.2 and are using earlier approaches to change the
default file associations (for example, through Windows Explorer), you
might encounter problems. This topic has generated a lively discussion
among the SAS Deployment Support Community! The recommended
method for changing file type associations is described in this paper. Also
discussed are ways to troubleshoot problems that might surface during this
process.
Opinion Mining and Geo-positioning of Textual
Feedback from Professional Drivers
2:00 p.m.
Mantosh Kumar Sarkar, Oklahoma State University
Goutam Chakraborty, Oklahoma State University
Paper 500-2013
While many companies collect feedback from their customers via mobile
applications, they often restrict their analysis to numeric data and ignore
analyzing customer feedback and sentiments from textual data. In this
paper, we analyze customer feedback by professional drivers sent via a
mobile app. We demonstrate how SAS® Text Miner can be used to
automatically generate and summarize topics from positive and negative
feedbacks. In addition, we demonstrate how SAS® Sentiment Analysis
studio can be used to build rules to predict customers’ sentiments
automatically so that experts’ time can be used for more strategic purposes.
Finally, we show how feedback with positive and negative sentiments can
be geo-positioned on the U.S. map via JMP® scripts to provide a better
visualization of sentiment distribution.
Foundations and Fundamentals — Room 2008
10:30 a.m.
Quick Hits: My Favorite SAS® Tricks
Marje Fecht, Prowerk Consulting
Paper 114-2013
Are you time-poor and code-heavy? It's easy to get into a rut with your SAS®
code, and it can be time-consuming to spend your time learning and
implementing improved techniques. This presentation is designed to share
quick improvements that take 5 minutes to learn and about the same time
to implement. The quick hits are applicable across versions of SAS and
require only Base SAS® knowledge. Included are: - little-known functions
that get rid of messy coding - simple macro tricks - dynamic conditional
logic - data summarization tips to reduce data and processing - generation
data sets to improve data access and rollback - testing tips
A Day in the Life of Data - Part 1
Brian Bee, The Knowledge Warehouse Ltd
(Invited) Paper 116-2013
As a new SAS® programmer, you may be overwhelmed with the variety of
tricks and techniques that you see from experienced SAS programmers; as
you try to piece together some of these techniques you get frustrated and
perhaps confused because the data showing these techniques are
inconsistent. That is, you read several papers and each uses different data.
This series of four papers is different. They will step you through several
techniques but all four papers will be using the same data. The authors will
show how value is added to the data at each of the four major steps: Input,
Data Manipulation, Data and Program Management, and Graphics and
Reporting.
3:00 p.m.
A Day in the Life of Data - Part 2
Harry Droogendyk, Stratia Consulting Inc.
(Invited) Paper 117-2013
As a new SAS® programmer, you may be overwhelmed with the variety of
tricks and techniques that you see from experienced SAS programmers; as
you try to piece together some of these techniques you get frustrated and
perhaps confused because the data showing these techniques are
inconsistent. That is, you read several papers and each uses different data.
This series of four papers is different. They will step you through several
techniques but all four papers will be using the same data. The authors will
show how value is added to the data at each of the four major steps: Input,
Data Manipulation, Data and Program Management, and Graphics and
Reporting.
4:00 p.m.
A Day in the Life of Data - Part 3
Peter Crawford, Crawford Software Consultancy Limited
(Invited) Paper 118-2013
As a new SAS® programmer, you may be overwhelmed with the variety of
tricks and techniques that you see from experienced SAS programmers; as
you try to piece together some of these techniques you get frustrated and
perhaps confused because the data showing these techniques are
inconsistent. That is, you read several papers and each uses different data.
10
www.sasglobalforum.org/2013
This series of four papers is different. They will step you through several
techniques but all four papers will be using the same data. The authors will
show how value is added to the data at each of the four major steps: Input,
Data Manipulation, Data and Program Management, and Graphics and
Reporting.
5:00 p.m.
A Day in the Life of Data: Part 4 - Graphics and Reporting
Sanjay Matange, SAS
Paper 119-2013
Although we often work in our own department with little contact with
others, whether they’re on the next floor or halfway around the world,
everyone has an impact on each other. In this four-part series, each paper
examines one of the four major aspects of SAS® Data Management: initial
data input; data manipulation; data and program management; and
graphics and reporting. Each paper teaches some fundamental skills and
shows how each step adds value to the data.
Hands-on Workshops — Room 2011
Hands-on Workshops — Room 2024
10:30 a.m.
Adding New Rows in the ADaM Basic Data Structure:
When and How
Mario Widel, Roche Molecular Systems
Sandra Minjoe, Octagon Research Solutions
(Invited) Paper 137-2013
The ADaM (Analysis Data Model) BDS (Basic Data Structure) has specific
rules to follow when adding columns or rows. Because there are limitations
to what can be added as a column, much of our derived content must be
added as rows. This HOW uses a Vital Signs example, demonstrating the
common BDS need of adding analysis parameters and visits. Attendees will
use a general specification and mock-up to create metadata content that
can be used for both a detailed specification and within a define document.
The resulting content will include variable-level metadata, parameter-level
metadata, and SAS® code snippets. This is an intermediate-level HOW.
Attendees are expected to be familiar with the analysis needs of clinical
trials, CDISC, and submissions to FDA.
10:30 a.m.
Hands-on Workshops — Room 2011
SAS® Workshop: SAS® Add-In for Microsoft Office 5.1
11:30 a.m.
Paper 520-2013
SAS® Workshop: Creating SAS® Stored Processes
Eric Rossland, SAS
This workshop provides hands-on experience using the SAS® Add-In for
Microsoft Office. Workshop participants will:
• access and analyze data
• create reports
• use the SAS add-in Quick Start Tools
Eric Rossland, SAS
Paper 521-2013
This workshop provides hands-on experience creating SAS® Stored
Processes. Workshop participants will:
• use SAS® Enterprise Guide® to access and analyze data
• create stored processes which can be shared across the organization
Hands-on Workshops — Room 2020
10:30 a.m.
The Armchair Quarterback: Writing SAS® Code for the
Perfect Pivot (Table, That Is)
Peter Eberhardt, Fernwood Consulting Group Inc
(Invited) Paper 136-2013
“Can I have that in Excel?” This is a request that makes many of us shudder.
Now your boss has discovered Microsoft Excel pivot tables. Unfortunately,
he has not discovered how to make them. So you get to extract the data,
massage the data, put the data into Excel, and then spend hours rebuilding
pivot tables every time the corporate data are refreshed. In this workshop,
you learn to be the armchair quarterback and build pivot tables without
leaving the comfort of your SAS® environment. You learn the basics of Excel
pivot tables and, through a series of exercises, how to augment basic pivot
tables first in Excel, and then using SAS. No prior knowledge of Excel pivot
tables is required.
• access the new stored process from the SAS® Add-In for Microsoft Office
2:00 p.m.
SAS® Workshop: SAS® Data Integration Studio Basics
Kari Richardson, SAS
Paper 522-2013
This workshop provides hands-on experience using SAS Data Integration
Studio to construct tables for a data warehouse. Workshop participants will:
• define and access source data
• define and load target data
• work with basic data cleansing
www.sasglobalforum.org/2013
11
Hands-on Workshops — Room 2020
2:00 p.m.
SAS® Enterprise Guide® 5.1: A Powerful Environment for
Programmers, Too!
Marje Fecht, Prowerk Consulting
Rupinder Dhillon, Dhillon Consulting Inc
(Invited) Paper 138-2013
Have you been programming in SAS® for a while and just aren't sure how
SAS® Enterprise Guide® can help you? This presentation demonstrates how
SAS programmers can use SAS Enterprise Guide 5.1 as their primary
interface to SAS, while maintaining the flexibility of writing their own
customized code. We explore:
• navigating and customizing the SAS Enterprise Guide environment
• using SAS Enterprise Guide to access existing programs and enhance
processing
• exploiting the enhanced development environment including syntax
completion and built-in function help
• using SAS® Code Analyzer, Report Builder, and Document Builder
• adding Project Parameters to generalize the usability of programs and
processes
• leveraging built-in capabilities available in SAS Enterprise Guide to
further enhance the information you deliver
• Review / create a SAS Data Integration Studio job that will execute the
uploaded data jobs on the DataFlux Data Management Server
4:00 p.m.
SAS® Workshop: SAS® Visual Analytics 6.1
Eric Rossland, SAS
Paper 524-2013
This workshop provides hands-on experience with SAS® Visual Analytics.
Workshop participants will:
• explore data with SAS® Visual Analytics Explorer
• design reports with SAS® Visual Analytics Designer
5:00 p.m.
SAS® Workshop: DataFlux® Data Management Studio
Basics
Kari Richardson, SAS
Paper 525-2013
This workshop provides hands-on experience using DataFlux® Data
Management Studio to profile then cleanse data. Workshop participants
will:
• learn to navigate DataFlux® Data Management Studio
• define and run a data profile
• define and run a data job
Hands-on Workshops — Room 2024
2:00 p.m.
IT Management — Salon 10,11, 12
Using PROC FCMP to the Fullest: Getting Started and
Doing More
1:30 p.m.
Art Carpenter, CA Occidental Consultants
(Invited) Paper 139-2013
The FCMP procedure is used to create user-defined functions. Many users
have yet to tackle this fairly new procedure, while others have attempted to
use only its simplest options. As with many tools within SAS®, the true value
of this procedure is appreciated only after the user has started to learn and
use it. The basics can quickly be mastered, and this allows the user to move
forward to explore some of the more interesting and powerful aspects of
the FCMP procedure. The use of PROC FCMP should not be limited to the
advanced SAS user. Even those fairly new to SAS should be able to
appreciate the value of user-defined functions.
Manage Your Data as a Strategic Asset
Khaled Ghadban, Canada Post
Richard Beaver, United Natural Foods, Inc
Bill Ford, Vail Resorts
(Invited) Paper 506-2013
3:00 p.m.
Searching for Business Value in Big Data with Hadoop
Mike Olson, Cloudera
Paul Kent, SAS
Gavin Day, SAS
Hands-on Workshops — Room 2011
(Invited) Paper 507-2013
3:00 p.m.
While some well-resourced organizations can simply throw technical talent
at uncovering the value in their big data, others struggle to find analytic
technology that takes full advantage of the richness and scale of the
Hadoop ecosystem. Join industry thought leaders from Cloudera, Intel and
SAS for a discussion of how the Hadoop community is using analytics to
derive critical insights that drive significant business impact from their big
data assets
SAS® Workshop: SAS® Data Integration Studio Advanced
Kari Richardson, SAS
Paper 523-2013
This workshop provides hands-on experience using a combination of
DataFlux Data Management Studio and SAS® Data Integration Studio.
Workshop participants will:
• Review two DataFlux Data Management Studio data jobs
• Upload the DataFlux Data Management Studio data jobs to the DataFlux
Data Management Server
12
www.sasglobalforum.org/2013
4:00 p.m.
3:30 p.m.
How IT Completes the Big Data Puzzle with Hadoop
Is the Legend in Your SAS/Graph® Output Still Telling the
Right Story?
Mike Olson, Cloudera
Pauline Nist, Intel
Paul Kent, SAS
Gavin Day, SAS
(Invited) Paper 508-2013
It's easy to become overwhelmed by the increasing volume, velocity and
variety of big data – and miss the value that it holds to uncovering
profitable insights and answering complex questions. So what's the missing
piece to solving the big data puzzle? Hadoop. IT organizations are rapidly
leveraging Hadoop to quickly derive a more complete picture and analysis
of all their data. Before you can get value from your data, it has to be well
organized, managed and governed. Thought leaders from Cloudera, Intel
and SAS will share key insights on how IT can solve the big data puzzle with
Hadoop
Pharma and Health Care — Room 2000
2:00 p.m.
Automated and Customized Reports as a Single Image
File Using Graph Template Language (GTL): A Case Study
of Benchmarking Reports in Medical Research
Monarch Shah, ICON Clinical Research
Ginny Lai, ICON Late Phase & Outcomes Research
Eric Elkin, ICON
Paper 495-2013
Site benchmarking reports give us an overview of demographic, clinical,
and disease characteristics for the individual site with comparison to the
study as a whole. A solution was needed for on-going reporting to over 250
study sites. The report needed to be concise and present data in both
tables and figures. (This objective could also arise, for example, in
comparing each store’s performance to the entire chain or each classroom’s
performance to the school district.) However, creating and combining
tables and figures into a document can be challenging. Graph Template
Language (GTL) provides a powerful alternative to customize and automate
benchmarking reports. This paper will focus on using GTL to create panels
comprised of descriptive tables and multiple graphs into a single image file.
2:30 p.m.
Patient Profile Graphs Using SAS®
Sanjay Matange, SAS
Paper 160-2013
Patient profiles provide information on a specific subject participating in a
study. The report includes relevant data for a subject that can help correlate
adverse events to concomitant medications and other significant events as
a narrative or a visual report. This presentation covers the creation of the
graphs useful for visual reports from CDISC data. It includes a graph of the
adverse events by time and severity, graphs of concomitant medications,
vital signs and labs. All the graphs are plotted on a uniform timeline, so the
adverse events can be correlated correctly with the concomitant
medications, vital signs and labs. These graphs can be easily incorporated
with the rest of the demographic and personal data of the individual
patient in a report.
Alice Cheng, Chiltern Inc.
Justina Flavin, self employed
(Invited) Paper 161-2013
In clinical studies, researchers are often interested in the effect of treatment
over time for multiple treatments or dosing groups. Usually, in a graphical
report, the measurement of treatment effect is on the vertical axis and a
second factor, such as time or visit, on the horizontal axis. Multiple lines are
displayed in the same figure; each line represents a third factor, such as
treatment or dosing group. It is critical that the line appearance (color,
symbol and style) is consistent throughout the entire clinical report as well
as across clinical reports from related studies.
4:30 p.m.
Variance Partition: My Mission and Ambition Come to
Fruition
Brenda Beaty, University of Colorado
L. Miriam Dickinson, University of Colorado
Paper 162-2013
In medical research, we are often interested in understanding the complex
interplay of variables with one or more clinical outcomes. Because our
bodies are always in motion, simply viewing a snapshot of data in time is
sub-optimal. Longitudinal data gives us the advantage of modeling 'reallife' time-dependent variables and outcomes. This paper is an exploration
of one such project. In this paper, we first familiarize ourselves with a study
of the relationship of diabetic nephropathy and blood pressure measured
longitudinally. We then explore a number of ways to model the data, with
the final goal of using time-varying covariates to model the illness path, as
well as the ultimate outcome, thereby getting complete partitioning of the
variance.
5:00 p.m.
Quantile Regression in Pharmaceutical Marketing
Research
George Mu, IMS health Inc
Paper 163-2013
In pharmaceutical marketing research, the heterogeneity in healthcare data
presents lots of challenges to researchers. Managers have a difficult time
getting comprehensive market pictures from simple equations that
generally fit all individuals. Quantile regression offers an efficient and
robust way to tease out the different patterns existing in the healthcare
market. This paper demonstrates the value of applying quantile regression
to solve pharmaceutical marketing research problems. We illustrate the
methodology by using SAS® QUANTREG and QUANTLIFE procedures to
compare physicians’ new product uptake patterns; to find influential drivers
in patient medication compliance; and to help in the design of clinical trials
for patient selections. The results from these empirical examples show
quantile regressions provide more market insights than other commonly
used methodologies.
www.sasglobalforum.org/2013
13
5:30 p.m.
Using PROC GENMOD to Investigate Drug Interactions:
Beta Blockers and Beta Agonists and Their Effect on
Hospital Admissions
Hui Fen Tan, Columbia University
Ronald Low, New York City Health and Hospitals
Corporation
Shunsuke Ito, New York City Health and Hospitals
Corporation
Raymond Gregory, New York City Health and Hospitals
Corporation
Vann Dunn, New York City Health and Hospitals Corporation
for a group of this nature to be successful in national and global
organizations. It also reviews technologies that could be beneficial for
bridging the communication gap in a user group of this type.
Planning and Support — Room 3016
3:00 p.m.
Branding Yourself Online
Kirsten Hamstra, SAS
Shelly Goodin, SAS
Paper 184-2013
Every year, more than half a million adverse reactions to drugs are reported
to the FDA. This paper is a real-world, large-scale review of beta blockers
and beta agonist usage. We use New York City public hospitals’ records to
investigate whether interactions of beta blockers and beta agonists are
associated with adverse medical outcomes such as increased hospital visits,
a common indicator of health care quality. The GENMOD procedure in SAS®
provides a variety of count data models, including Poisson regression and
negative binomial regression. We find that patients on “non-clinical trials
use” of beta blockers and beta agonists, older patients, and patients with
history of COPD, CAD, and pneumonia tend to have higher hospital visit
rates.
Your online reputation matters. Whether you’re using social media for
professional or personal reasons, it’s important to understand and control
your public persona. For those looking to further your career, build your
business, or enter the workforce, you can maximize your positive exposure
by knowing where and how to engage online. This presentation will
highlight some of the best practices for online engagement, provide
suggestions for where to engage, and showcase some examples. To anyone
engaging in social media who wants to better understand the impact of
their contributions and control their online reputation, this presentation is
for you. Takeaways: · Craft an incredible bio · Harness the power of SEO ·
Strengthen your online reputation · Engage—where and how to do it ·
Make meaningful connections Audience: · Job seekers · Consultants ·
Students
Planning and Support — Room 2010
Planning and Support — Room 2010
2:00 p.m.
4:00 p.m.
SAS® Skill Learning and Certification Preparation in a
Graduate School Setting
Coaching SAS® Beginner Programmers: Common
Problems and Some Solutions
Paper 164-2013
Christine Bonney, University of Pennsylvania
Michael Keith, Jr., University of Pennsylvania
Paper 182-2013
A semester-long course was created with the goal of teaching graduate
students SAS® programming skills and to prepare them to take the SAS®
Certified Base Programming for SAS®9 exam. Course activities and materials
include: weekly lectures; in-class labs; take-home problem sets; virtual
(online) labs; assigned readings from the “SAS® Certification Prep Guide:
Base Programming for SAS®9”; midterm and final exams; and access to SAS®
OnDemand for Academics. This paper covers the details of the course
development and design, as well as preliminary results from the course and
plans for future developments.
2:30 p.m.
Considerations for Creating an In-House SAS® User
Group in a Geographically Disbursed Organization
Stefanie Reay, Qualex Consulting Services, Inc.
Paper 183-2013
This presentation will review considerations for creating in-house SAS® user
groups in geographically disbursed organizations, in which in-person user
group meetings are not cost-effective or not feasible, but for which an inhouse SAS user group would still be beneficial. It defines in-house SAS user
groups, and overviews the resources available from SAS for starting and
continuing an in-house SAS user group. It discusses benefits and challenges
of starting/maintaining an in-house SAS user group, options for
organizational structures of in-house SAS user groups, and unique needs
14
www.sasglobalforum.org/2013
Peter Timusk, Statistics Canada
Paper 185-2013
This paper will present a number of problems SAS® beginner programmers
encounter when first writing SAS programs. The paper will cover three
cases and show how pointing out patterns to beginner programmers will
aid them in avoiding errors in their SAS code.
4:30 p.m.
A CareerView Mirror: Another Perspective on Your Work
and Career Planning
Bill Donovan, Ockham Source
Paper 186-2013
Career planning in today’s tumultuous job market place requires a more
rigorous and disciplined approach that must begin with each individual
tracking and evaluating distinctive skills and experiences. With an emphasis
on the SAS® professional and the career track unique to the programmers'
challenges, this paper is designed to set the stage for professional reflection
and career planning. The ability to organize and inventory your entire
career-related experiences is the foundation of a solid plan. The catalog of
your work assignments and functional responsibilities creates a reflection
of your efforts in your career to date. All of this helps to build your
CareerView Mirror, which provides another perspective on your work and
career planning.
5:00 p.m.
2:00 p.m.
Gotchas: Hidden Workplace and Career Traps to Avoid
SAS® Essentials: Maximize the Efficiency of Your Most
Basic Users
Steve Noga, Rho
Paper 187-2013
Being successful at your job takes more than just completing your tasks
accurately and on time. There are hidden holes everywhere, some deeper
than others, that must be navigated; yet no map exists for you to follow.
Most companies have a set of stated policies or rules that their employees
are expected to follow, but what about the unstated ones that may have an
effect on how fast or how far you advance within the company? Hidden
traps also exist along the way of your career path. This panel discussion will
highlight some “gotchas” of which you should be aware and ways to keep
from falling into the holes.
Poster and Video Presentations — SAS Support
and Demo Area
2:00 p.m.
Create a Nomogram with SGPlot
Julie Kezik, Yale University
Melissa Hill, Yale University
Paper 196-2013
If programming and research assistants were taught SAS® essentials, job
efficiency could be maximized with the ability to use SAS as a tool to do
their own preparatory work for assigned tasks. This paper summarizes a
supplemental training program which teaches basic SAS programming
skills to enable support staff to be more independent.
2:00 p.m.
Using CALL SYMPUT to Generate Dynamic Columns in
Reports
Sai Ma, pharmanet-i3
Suwen Li, Everest Clinical Research Services, Inc.
Regan Li, Hoffmann-La Roche Limited
Bob Lan, Everest Clinical Research Services, Inc.
Cynthia Loman, Genomic Health Inc
Paper 198-2013
The nomogram that I created shows the relative values of four predictors
from a logistic model along with a line showing cumulative model score
and cumulative model risk. I have programmed it with PROC SGPLOT and
used natural splines for one of the variables in my model. I used prostate
cancer data for my example.
When creating reports, we often want to make the report respond
dynamically to data. If the headers and the number of columns in the report
are unknown, it is helpful when they change dynamically depending on the
data. As a powerful SAS® procedure, PROC TABULATE can produce dynamic
results in most cases. This paper describes how to use the CALL SYMPUT
routine and PROC REPORT to generate dynamic columns in reports in cases
where PROC TABULATE does not yield the desired results.
Paper 194-2013
2:00 p.m.
Using SAS to Create Code for Current Triage Systems
during Chemical Incidents
Abbas Tavakoli, University of South Carolina
Erik Svendsen, University of tulane
Jean Craig, MUSC
Joan Culley, University of South Carolina
2:00 p.m.
From SDTM to ADaM
Sai Ma, pharmanet-i3
Suwen Li, Everest Clinical Research Services, Inc.
Regan Li, Hoffmann-La Roche Limited
Bob Lan, Everest Clinical Research Services, Inc.
Paper 195-2013
Paper 199-2013
Chemical incidents involving irritant chemicals such as chlorine pose a
significant threat to life and require rapid assessment. This paper used the
first outcomes-level study (R21 NIH) involving an actual mass casualty
chemical incident to create code for four triage systems (CBRN, SALT,
START, and ESI). Data used for this paper, which come from six datasets
collected by the project team from a 60-ton railroad chlorine leak in 2005 in
Graniteville, South Carolina, include patient demographics, exposure
estimates, symptoms, outcome categories, and physiological
measurements. Data collected for approximately 900 victims of the chlorine
leak were merged to generate a research dataset. SAS® 9.2 was used to
create code from logic to mimic the triage decision tree, yielding
classifications for each system.
The use of SDTM and ADaM standards are highly desirable in FDA
guidances. More and more sponsors submit both of these standards to
regulatory authorities. When SDTM data sets are more common, ADaM is
usually derived from SDTM. However, the SDTM distinctive data structure
causes problems when deriving ADaM data. This paper describes problems
encountered when deriving ADaM data and provides resolutions and
examples.
2:00 p.m.
Exploring the PROC SQL _METHOD Option
Charlie Shipp, Consider Consulting, Inc.
Kirk Paul Lafler, Software Intelligence Corporation
Paper 200-2013
The SQL procedure has powerful options for users to take advantage of.
This presentation explores the fully supported _METHOD option as an
applications development and tuning tool. Attendees learn how to use this
powerful option to better understand and control how a query processes.
www.sasglobalforum.org/2013
15
2:00 p.m.
2:00 p.m.
A Practical Approach to Creating Define.XML by Using
SDTM Specifications and Excel functions
A Simple Macro to Minimize Data Set Size
Paper 201-2013
Define.xml (Case Report Tabulation Data Definition Specification) is a part
of new drug submission required by the FDA. Clinical SAS® programmers
usually use SAS programming [1, 2, 3, 4, 5] to generate the code of
Define.xml as described in the CDISC Case Report Tabulation Data
Definition Specification (define.xml) V1.0.0 [6]. This paper illustrates the
process of using SDTM specifications and Excel functions to generate the
code of Define.xml in an easy and straightforward way.
Whenever you submit either SDTM or ADaM data sets to FDA, if any SAS®
data set is great than 1 GB in size, FDA will ask you to split the data set. In
fact, since the length of a variable affects both the amount of disk space
used and the number of I/O operations required to read and write the data
set, resizing text columns to fit the longest value within the column is
applicable to every field that uses SAS data sets in their business. To help
save resources and improve data mining efficiency, this paper discusses a
simple macro to minimize the size of a SAS data set.
2:00 p.m.
2:00 p.m.
Extending the Power of Base SAS® with Microsoft Excel
Basel II Advanced IRB in Commercial Banking: Quantify
the Borrower and Guarantor by Two-Step Scoring Model
Amos Shu, Endo Pharmaceuticals
Shilpa Khambhati, Mathematca Policy Research Inc.
Paper 203-2013
The SAS Macro Language is an invaluable SAS tool that can be used for
iterative SAS data processing, eliminating redundancy in SAS code. Using
the SAS Macro Language with Microsoft Excel makes programming tasks
even easier. This paper describes using the SAS Macro Language and
Microsoft Excel to automatically generate customized reports. The
proposed method uses Excel macros to drive SAS macros without having to
open SAS programs and manually upgrade parameters specific for each
site’s data when the data becomes available. The process eliminates
manually editing SAS programs and improves data quality by reducing
programming error and program maintenance time.
2:00 p.m.
Using LinkedIn to Find Your Next SAS® Job
Tricia Aanderud, And Data Inc
Paper 204-2013
LinkedIn is fast becoming a great place for SAS® recruiters and SAS
candidates to meet. If you are looking for a job, this poster provides some
tips to spiff up your LinkedIn profile to get the SAS programming job of
your dreams.
2:00 p.m.
Selection Group Prompts with SAS® Stored Processes:
More Power, Less Programming
Tricia Aanderud, And Data Inc
Angela Hall, SAS
Paper 205-2013
Many programmers either do not know or understand how to use the
selection group prompts to make advanced stored processes a little easier
to manage. Many times, end users have these crazy requirements and a
programmer can use the selection group prompt instead of writing 10
different stored processes.
16
www.sasglobalforum.org/2013
Amos Shu, Endo Pharmaceuticals
Paper 206-2013
Hengrui Qu, Citi Group Inc.
juan zhao, Citi Group Inc
Paper 207-2013
For public companies, the probability of default usually adopts well-known
structural and reduced form credit risk models. However, in commercial
lending, there are large portfolios of unlisted companies, which could not
use these two approaches. Furthermore, privately held companies
commonly get a guarantor to enhance their credibility during loan
application. Unlike the single logistic model used for retail credit risk
analysis, two- step credit scoring models could be used to quantify both
borrower and guarantor's risk exposed to unlisted companies based on the
limited information maximum likelihood. This paper will focus on how to
quantify the risk for commercial borrowers with guaranty by two-step
scoring model, which provides Basel II advanced IRB risk measure: the PD
for the commercial customer and transaction.
2:00 p.m.
A Unique Approach to Create Custom Reports By
Leveraging the Strengths of SAS® and Excel
Amy Overby Wilkerson, RTI International
Brett Anderson, RTI International
Barbara Bibb, RTI International
Mai Nguyen, RTI International
Paper 208-2013
Survey projects often require custom reports to allow project staff to
monitor production as well as various statistics from the collected data. At
RTI, we've come up with a unique approach for creating custom reports for
our projects by leveraging the strengths of SAS® and Excel. In SAS, we use
PROC SQL to select and when necessary aggregate data. After processing
the data in SAS, results are sent to Excel for reporting and graphics. In our
paper, we will present a few sample reports, program codes and the
detailed explanations of how these reports were created.
2:00 p.m.
Working with a Large Pharmacy Database: Hash and
Conquer.
David Izrael, Abt Associates
Paper 209-2013
Working with a large pharmacy database means having to process - merge,
sort, and summarize - hundreds of millions of observations. By themselves,
traditional methods of processing can lead to prohibitive data processing
times that endanger deadlines. The hash object is the fastest and most
versatile method in the SAS® system of substantially accelerating the
processing. In our paper, we apply hash methods to a routine lookup
function where one needs to merge the kernel pharmacy database with its
satellites. We also present comparatively new nontraditional features of the
hash object, such as handling duplicate keys and finding frequency
counters. At the same time, we underscore the necessity of traditional sortand-merge methods, but suggest that they be used carefully.
2:00 p.m.
With a Trace: Making Procedural Output and ODS
Output Objects Work for You
Louise Hadden, Abt Associates Inc.
(Invited) Paper 210-2013
The Output Delivery System (ODS) delivers what used to be printed output
in many convenient forms. What most of us don't realize is that "printed
output" from procedures (whether the destination is PDF, RTF, or HTML) is
the result of SAS® packaging a collection of items that come out of a
procedure that most people want to see in a predefined order (aka
template). With tools such as ODS TRACE, PROC CONTENTS, and PROC
PRINT, this paper explores the many buried treasures of procedural output
and ODS output objects and demonstrates how to use these objects to get
exactly the information that is needed, in exactly the format wanted.
2:00 p.m.
Analyzing the Safewalk Program with SAS®: Saving
Shelter Dogs One Walk at a Time
Louise Hadden, Abt Associates Inc.
Terri Bright, MSPCA Boston
footnotes, ODS text fields and tabular output; and add custom "fills" to SAS/
GRAPH® maps and graphs. Some possible uses of custom images include a
company logo embedded in SAS output, graphic displays of positive or
negative outcomes, and watermarks containing "draft" or "confidential".
The SAS code to accomplish all these potential uses, and more, will be
shown.
2:00 p.m.
Weighted Sequential Hot Deck Imputation: SAS® Macro
vs. the SUDAAN PROC HOTDECK
David Izrael, Abt Associates
Michael Battaglia, Abt Associates Inc
Paper 213-2013
Item non-response is a challenge faced by all surveys. Item non-response
occurs when a respondent skips over a question, refuses to answer a
question, or does not know the answer to a question. Hot deck imputation
is one of the primary imputation tools used by survey statisticians. Recently,
a new competitor in the field of Weighted Sequential Hotdeck Imputation
has arrived: PROC HOTDECK of SUDAAN, version 10. We compared the
results of imputation using the new procedure with the results of the
Hotdeck SAS® Macro with respect to: a) how close the post-imputation
weighted distributions and standard errors of the estimates are to those of
the item respondent data; b) whether there is a difference in the number of
times donors contribute to the imputation.
2:00 p.m.
Creating ZIP Code-Level Maps with SAS®
Barbara Okerson, WellPoint
(Invited) Paper 214-2013
SAS®, SAS/GRAPH®, and ODS graphics provide SAS programmers with the
tools to create professional and colorful maps. Provided with SAS/GRAPH
are boundary files for U.S. states and territories, as well as internal
boundaries at the county level. While much data and results can be
displayed at this level, often a higher degree of granularity is needed. The
U.S. Census Bureau provides ZIP code boundary files in ESRI shape file
format (.shp) by state for free download and import into SAS using SAS
PROC DATAIMPORT. This paper illustrates the use of these ZIP code
tabulation area (ZCTA) files with SAS to map data at a ZIP code level.
Example maps include choropleth, distance, and heat maps.
(Invited) Paper 211-2013
The MSPCA in Boston initiated the Safewalk Program in January 2009. This
program was designed to enrich the experience of shelter dogs by
providing training to volunteers and staff that allow dogs of varied
backgrounds and temperaments to be exercised safely, as well as
promoting behaviors encouraging adoption on the adoption floor. A data
extract from the MSPCA's Chameleon data base was analyzed using
multiple SAS® procedures in SAS/STAT®. This paper will demonstrate how
SAS analysis, output, and statistical graphs allowed us to assess the effects
of the Safewalk Program and which populations it most affected.
2:00 p.m.
Behind the Scenes with SAS®: Using Custom Graphics in
SAS Output
Louise Hadden, Abt Associates Inc.
(Invited) Paper 212-2013
SAS® provides many opportunities to add customized images to SAS ODS
output. This presentation will demonstrate various ways to add custom
backgrounds to tabular and graphic output; add custom images to titles,
2:00 p.m.
A Case Application of Propensity Score Matching in MTM
Outcomes Evaluation at Retail Pharmacy
Michael Taitel, Walgreens
Zhongwen Huang, Walgreens
Youbei Lou, Walgreens
Paper 215-2013
Propensity score matching approaches in outcomes analysis are often used
to reduce the potential bias in observational studies. The process includes
propensity score estimation, matching and evaluation. This paper presents
a case application in Outcome Evaluation of Medication Therapy
Management at Retail Pharmacy. Baseline outcome metrics, which do not
appear in the propensity score estimation model, were checked for balance
later to detect if there are any important covariates that affect both
treatment and outcomes have been neglected. In addition, by
appropriately selecting variables in retail pharmacy environment, one
matching for multiple outcomes analysis, which works as a pseudo
randomization study, can improve efficiency.
www.sasglobalforum.org/2013
17
2:00 p.m.
2:00 p.m.
Using SAS® to Expand the Application of Standard
Measures and Guide Statistical Explorations: Creating
Healthy Eating Index Scores Using Nutrition Data
System for Research Output
Propensity Score-Based Analysis of Short-Term
Complications in Patients with Lumbar Discectomy in
the ACS-NSQIP Database
David Ludwig, University of Miami
David Landy, University of MIami Miller School of Medicine
Joy Kurtz, Univ. of Miami
Tracie Miller, University of Miami
Paper 216-2013
We created a SAS® program to calculate a measure of diet quality, the
Healthy Eating Index (HEI, http://www.cnpp.usda.gov/
HealthyEatingIndex.htm), using output from a widely applied dietary
software package, Nutrition Data System for Research (NDSR, http://
www.ncc.umn.edu/products/ndsr.html). Currently, application of the HEI in
research and clinical assessment is limited by the challenges posed in
calculating the HEI using the highly complex and detailed NDSR output.
The SAS program extracts the required NDSR output files and then
calculates the combination of algebraic manipulations and logical
statements to obtain HEI scores. We also offer suggestions for increasing
usability, such as with the %INCLUDE statement, and show how the
program can be used to explore related statistical issues such as reliability
via PROC MIXED.
2:00 p.m.
A SAS® Macro for Generating a Set of All Possible
Samples with Unequal Probabilities without
Replacement
Alan Silva, Universidade de Brasilia
Paper 217-2013
This paper considers listing all possible samples of size n with unequal
probabilities without replacement in order to find the sample distribution.
The main application of that is to estimate the Horvitz-Thompson (HT)
estimator and possibly to know the shape of its sample distribution to
construct confidence intervals. The algorithm computes all possible
samples of the population, in contrast with PROC SURVEYSELECT which
generates any samples of size n, but not all possible samples, and at the
end it is possible to plot the sample distribution of the estimator. The
equations are encoded in a SAS/IML® macro and the graphics are made
using PROC GPLOT.
2:00 p.m.
Here Is How We Do It: Teaching SAS® at Community
Colleges
Meili Xu, Ohlone College
Paper 218-2013
Data is everywhere today, and SAS® programming skills are in high
demand. Providing community college students with SAS skills is extremely
valuable in preparing them for real-world job positions right after taking
the classes. In this paper, we will describe our experience and approaches
to teaching SAS to our students at Ohlone College. With the paper
presentation at the conference, we wish to instigate a dialogue among
other educators teaching SAS to share ideas and resources so that we may
all better equip students with strong SAS skills that will serve them well in
their future careers. At the workshop, you may also have an opportunity to
gain some hands-on experience on basic SAS procedures.
18
www.sasglobalforum.org/2013
Yubo Gao, University of Iowa
Paper 220-2013
Lumbar discectomy is the most common spinal procedure performed, and
it can be done on an outpatient basis. In this study, we want to compare the
incidence of complications in patients undergoing single-level lumbar
discectomy between the inpatient and outpatient settings, to determine
baseline 30-day complication rates, and to identify independent risk factors
for complications. To achieve those, patients undergoing lumbar
discectomy between 2005 and 2010 were selected from the ACS-NSQIP
database, based on a single primary CPT code. Thirty-day post-operative
complications and pre-operative patient characteristics were identified and
compared. Propensity score matching and multivariate logistic regression
analysis were used to adjust for selection bias and identify predictors of 30day morbidity. All analyses are performed via SAS® software.
2:00 p.m.
Recovering SAS® User Group Proceedings for the SAS®
Community
Lex Jansen, lexjansen.com
Richard La Valley, Strategic Technology Solutions
Kirk Paul Lafler, Software Intelligence Corporation
Paper 221-2013
For many years, SAS® User Groups held conferences whose proceedings
were available only in print and only to those who attended or those who
knew that copies existed in the SAS Library in Cary, NC. Over the past
couple of years, there has been a project to digitize the printed proceedings
of SAS User Groups International, SAS European Users Group International,
NorthEast SAS Users Group, SouthEast SAS Users Group, Western Users of
SAS Software, South-Central SAS Users’ Group, MidWest SAS Users Group,
and the Pacific Northwest SAS Users Group. This paper provides an
overview of the project and the progress that has been made on this effort.
2:00 p.m.
GEN_ETA2: A SAS® Macro for Computing the Generalized
Eta-Squared Effect Size Associated with Analysis of
Variance Models
Patricia Rodriguez de Gil, University of South Florida
Thanh Pham, University of South Florida
Patrice Rasmussen, 5336 Clover Mist Drive
Jeanine Romano, University of South Florida
Yi-Hsin Chen, University of South Florida
Jeffrey Kromrey, University of South Florida
Paper 223-2013
Measures of effect size are recommended to communicate information on
the strength of relationships between variables. Such information
supplements the reject / fail-to-reject decision obtained in statistical
hypothesis testing. The choice of an effect size for ANOVA models can be
confusing because indices may differ depending on the research design as
well as the magnitude of the effect. Olejnik and Algina (2003) proposed the
generalized eta-squared effect size which is comparable across a wide
variety of research designs. This paper provides a SAS® macro for
computing the generalized eta-squared effect size associated with analysis
of variance models by utilizing data from PROC GLM ODS tables. The paper
provides the macro programming language, as well as results from an
executed example of the macro.
2:00 p.m.
2:00 p.m.
Anh Kellermann, USF
Aarti Bellara, University of South Florida
Patricia Rodriguez de Gil, University of South Florida
Diep Nguyen, University of South Florida
Eun Sook Kim, University of South Florida
Yi-Hsin Chen, University of South Florida
Jeffrey Kromrey, University of South Florida
Linking Laboratory Data To Submission Documents
Using SAS® Technologies
Dongmin Shen, Merck & Co, Inc
Paper 225-2013
Merck is a global pharmaceutical company and so the sources of our data
are global. Having the ability to link and transfer massive amounts of
analytical data from various data sources into submission documents in an
efficient and reproducible way is critical to producing successful regulatory
submissions. SAS® technologies have been used to create various solutions
ranging from data extractions, to data transformations, to documents
generated in support of simultaneous worldwide new drug applications.
2:00 p.m.
Linking Laboratory Data To Submission Documents
Using SAS® Technologies
Miu Ling Lau, Merck & Co.
Paper 225-2013
Merck is a global pharmaceutical company and so the sources of our data
are global. Having the ability to link and transfer massive amounts of
analytical data from various data sources into submission documents in an
efficient and reproducible way is critical to producing successful regulatory
submissions. SAS® technologies have been used to create various solutions
ranging from data extractions, to data transformations, to documents
generated in support of simultaneous worldwide new drug applications.
2:00 p.m.
Presenting Business Cases That Contain Complex,
Technical Information to a Varied Audience
Stephen Moore, US Census Bureau
Lori Guido, US Census Bureau
Paper 226-2013
The U.S. Census Bureau has a SAS® user base of approximately 2,600 users,
which the Software Application Branch (SADB) of the Applications Services
Division supports. In order to obtain the resources needed to provide the
support the users community needed, we had to figure out how to herd
cats. We had to gather information, enlist help from many sources, and get
everyone involved in the effort on the same level of understanding and
agreement. This paper describes the method SADB used to justify the
expansion of the Census SAS Support area from two to eight people. This
paper focuses on the following topics: the Census SAS support model, issue
definition, issue leveling, and communication strategy.
Variance Heterogeneity and Non-Normality: How the
SAS® TTEST Procedure Can Keep Us Honest
Paper 228-2013
The independent samples t-test is one of the most used tests for detecting
true mean differences. The SAS® System provides PROC TTEST, which is an
easy way to conduct a test for the difference between two population
means by assuming homogeneity of variance or avoiding it. However, the ttest and its alternatives (Satterthwaite's approximate test and conditional ttest) assume population normality; therefore, questions about the
performance of conditional testing when the assumption of normality is
not met remain. This paper describes previous research on preliminary tests
under the normality assumption, extends this research to the evaluation of
conditional testing to departures of normality, and provides guidance to
researchers on the proper use of this test with non-normal, heteroscedastic
population distributions.
2:00 p.m.
Here Comes Your File! File-Watcher Tool with
Automated SAS® Program Trigger
Rajbir Chadha, Cognizant Technology Solutions
Paper 229-2013
This paper talks about a file-watcher tool (UNIX Shell Script) that searches
for files and checks when they were last updated. Parameters to the filewatcher tool are supplied using a 'wrapper' script. Script is scheduled using
a 'CRON' scheduler in UNIX. Once file is found, SAS program is triggered. If
file is not found or not updated the script terminates. Tool sends out emails when files are available and when SAS program completes execution
or script terminates. In case of errors, users can refer to the file-watcher logs
at a location specified in the CRON file. The file-watcher tool reduces
average wait time and manual effort for users by automating most of the
process, allowing them to focus on other pressing tasks.
2:00 p.m.
Why the Bell Tolls 108 Times? Stepping through Time
with SAS®
Peter Eberhardt, Fernwood Consulting Group Inc
Paper 230-2013
For many SAS® programmers, the use of SAS date and datetime variables is
often very confusing. This paper addresses the problems that the most of
programmers have. It starts by looking at the basic underlying difference
between the data representation and the visual representation of date,
datetime, and time variables. From there, it discusses how to change data
representations into visual representations through the use of SAS formats.
The paper also discusses date arithmetic first by demonstrating the use of
simple arithmetic to increment dates; then by moving on to SAS functions
which create, extract, and manipulate SAS date variables. This paper is
introductory and focuses on new SAS programmers; however, some
advanced topics are also covered.
www.sasglobalforum.org/2013
19
2:00 p.m.
2:00 p.m.
Implementing CDISC, SDTM, and ADaM in a SAS®
Environment
A Preventive Approach for Automatic Checking of CDISC
ADaM Metadata to Detect Noncompliance
Pankaj Bhardwaj, Tata Consultancy Services
Paper 231-2013
Key challenges for regulatory bodies like FDA are non-standardized data
(almost 50% of the submissions) and its non-traceability. Reviewers cannot
streamline their review processes. A lot of work is happening in this
direction, and there is the expectation that all submissions will need to be
in standardized format by 2015 or so. This paper helps in building a
metadata-oriented, flexible, GUI-based solution for implementing the
CDISC and ADaM standards in a SAS® environment with following steps: 1.
Efficiently set up CDISC and ADaM metadata in SAS data sets, considering
important aspects like CDISC amendments and customization. 2. SAS
coding environment for handling legacy, ongoing and future trials. 3.
Generalized SAS validation codes for validation at the source data, SDTM,
and ADaMs level. 4. The submission deliverables.
2:00 p.m.
Weathering the Storm: Using Predictive Analytics to
Minimize Utility Outages
Mark Konya, Ameren Missouri
Kathy Ball, SAS
Paper 232-2013
Due to ever-increasing customer service expectations an ongoing
challenge for utilities is maintaining and improving the reliability of their
electric distribution systems. With significant numbers of transformers and
meters at risk of losing power during major storms, how can a Distribution
Engineer make sense of thousands of data points to prevent outages before
a storm occurs and, for customers whose power is interrupted during a
storm, restore service faster? Distribution Optimization equips utility
engineers and dispatchers to predict which assets will be affected by
storms while optimizing the placement of crews, thus decreasing outage
restoration times. Combining geospatial visualization with predictive
analytics, the predictive enterprise utility can shorten outages from weather
events and identify weak points in the electrical distribution system thus
preventing future outages.
2:00 p.m.
What Score Should Johnny Get? Missing_Items SAS®
Macro for Analyzing Missing Item Responses on
Summative Scales
Patricia Rodriguez de Gil, University of South Florida
Jeffrey Kromrey, University of South Florida
Paper 233-2013
Missing data are usually not the focus of any given study but researchers
frequently encounter missing data when conducting empirical research.
Missing data for Likert-type response scales, whose items are often
combined to make summative scales, are particularly problematic because
of the nature of the constructs typically measured, such as attitudes and
opinions. This paper provides a SAS® macro, written in SAS/IML® and SAS/
STAT®, for imputation of missing item responses that allows estimation of
person-level means or sums across items in the scale. Imputations are
obtained using multiple imputation (MI), single regression substitution
(SRS), relative mean substitution (RMS), and person mean substitution
(PMS). In addition, the results of a simulation study comparing the accuracy
and precision of the imputation methods are summarized.
20
www.sasglobalforum.org/2013
Min Chen, Vertex Pharmaceuticals, Inc.
Xiangchen Cui, Vertex Pharmaceuticals, Inc.
Tathabbai Pakalapati, CYTEL INC.
Paper 234-2013
The ADaM programming specification serves as the primary source for
ADaM programming, Define.xml, and reviewer guide. It should meet FDA
requirements and follow CDISC ADaM guidelines. OpenCDISC Validator is a
very useful tool to check the compliance with CDISC models. Sometimes it
is too late and/or costly to fix the errors identified by the tool. This paper
introduces a preventive approach to check metadata compliance with
ADaM guidelines at an earlier stage even before actual ADaM data set
programming thereby avoiding the waste of time and resources for
correction at a later stage. It also automatically ensures the consistency of
variable attributes between ADaM data sets and the define files, which
guarantees technical accuracy and operational efficiency.
2:00 p.m.
Building Traceability for End Points in Analysis Data Sets
Using SRCDOM, SRCVAR, and SRCSEQ Triplet
Xiangchen Cui, Vertex Pharmaceuticals, Inc.
Tathabbai Pakalapati, CYTEL INC.
Paper 235-2013
To be compliant with ADaM Implementation Guide V1.0, traceability
features should be incorporated to possible extent in analysis data sets.
SRCDOM, SRCVAR, and SRCSEQ triplet are used to establish data point
traceability in ADaM data sets. This paper provides various examples of
applying the triplet to establish traceability in efficacy ADaM data sets, and
shows the art of applying the triplet to different scenarios.
2:00 p.m.
Building Traceability for End Points in Analysis Data Sets
Using SRCDOM, SRCVAR, and SRCSEQ Triplet
Qunming Dong, Vertex
Tathabbai Pakalapati, CYTEL INC.
Paper 235-2013
To be compliant with ADaM Implementation Guide V1.0, traceability
features should be incorporated to possible extent in analysis data sets.
SRCDOM, SRCVAR, and SRCSEQ triplet are used to establish data point
traceability in ADaM data sets. This paper provides various examples of
applying the triplet to establish traceability in efficacy ADaM data sets, and
shows the art of applying the triplet to different scenarios.
2:00 p.m.
SAS® Admin's Best Friend - The Set-up and Usage of
RTRACE Option
Airaha Chelvakkanthan Manickam, Cognizant Technology
Solutions
Srikanth Thota, Cognizant Technology Solutions
Paper 236-2013
The SAS® license of any organization includes various SAS components
such as SAS/STAT®, SAS/GRAPH®, SAS/OR®, etc. How does a SAS
Administrator know how many of the licensed components are actively
used, how many SAS users are actively utilizing the server, and how many
SAS data sets are frequently referenced? These questions help a SAS
administrator make important decisions such as controlling SAS licenses,
removing inactive SAS users, purging long-time non-referenced SAS data
sets, etc. SAS provides a system parameter called RTRACE to answer these
questions. The goal of this paper is to explain the set-up of the RTRACE
parameter and to explain its usage in making the SAS administrator's life
easy. This paper is based on SAS® 9.2 running on AIX 6.1 operating system.
2:00 p.m.
SAS® Stored Processes Logging
Bhargav Achanta, Reata Pharmacueticals
Paper 237-2013
You and your colleagues work very hard to create stored processes and
deliver them to various departments in your organization to review the
reports on regular basis and on time. Have you ever wondered how many
of the reports you provide to the audience are actually being used? This
paper presents a neat way to identify who ran the stored processes and
what time they have run the stored processes by scanning all the stored
process server log files and generates a list report and a frequency report.
2:00 p.m.
Creating a Management-Friendly HTML Report Using
SAS® ODS Markup, Style Sheets, and JavaScript
Rosely Flam Zalcman, Center for Addiction and Mental
Health
Robert Mann, 33 Russell St
Paper 238-2013
SAS® ODS output of 127 individual tables using SAS tagsets are integrated
into a single portable active HTML file. Hyperlinks and embedded
JavaScript menus provide easy access to both Client Satisfaction statistics
(15 pages) and Service Provider analysis (112 pages).
2:00 p.m.
Implementation of Slowly Changing Dimension to Data
Warehouse to Manage Marketing Campaigns in Banks
2:00 p.m.
Feature Extraction and Rating of a Smartphone
Photosharing Application Using SAS® Sentiment
Analysis Studio
Goutam Chakraborty, Oklahoma State University
Siddhartha Reddy Mandati, Oklahoma State University
Anil Kumar Pantangi, Oklahoma State University
Sahithi Ravuri, Oklahoma State University
Paper 241-2013
Smartphone users often have to read many online reviews to find out about
an application's feature. Online reviews usually provide an overall numeric
rating using the Likert or semantic scale, but these reviews do not fully
reveal the sentiments of customers. In this paper, the website Google Play is
considered. Google Play is a dedicated portal for all Android paid and free
applications. SAS® Sentiment Analysis Studio is used to predict a review as
either positive or non-positive. To extract features of the application, builtin manual rules in the rule-based model are used. In this data, the rulebased model outperformed the statistical and hybrid model. The best
model helps categorize each review of the application by its features and its
rating.
2:00 p.m.
V is for Venn Diagrams
Kriss Harris, SAS Specialists
Paper 243-2013
Would you like to produce Venn diagrams easily? This poster shows how
you can produce stunning two-, three-, and four-way Venn diagrams by
using the SAS® Graph Template Language, in particular the DRAWOVAL and
DRAWTEXT statements. From my experience, Venn diagrams have typically
been created in the pharmaceutical industry by using Microsoft Excel and
PowerPoint. Excel is used to first count the numbers in each group, and
PowerPoint is used to generate the two- or three-way Venn diagrams. The
four-way Venn diagram is largely unheard of. When someone is brave
enough to tackle it manually, then working out the numbers that should go
in each of the 16 groups and inputting the right number into the right
group is usually done nervously!
Lihui Wang, SIngapore Management University
Michelle Cheong, Singapore Management University
Murphy Choy, Singapore Management University
2:00 p.m.
In this paper, we illustrate the concept of the slowly changing dimension
and how it can be utilized in an innovative manner in the data warehouse
of a bank to update and maintain campaign records of customers.
Paper 245-2013
Paper 239-2013
SAS® Grid Job Submission and Monitoring from the SAS®
Information Delivery Portal
Adolfo Lopez, Valence Helath
As part of the implementation of SAS® Grid Computing at Valence Health,
we realized that users would need a simple and straightforward way to
submit SAS® programs to the grid from their desktops. While the SAS® Grid
Manager Client Utility provides this functionality. it requires that additional
software be installed on the client computer and that the user be
comfortable with a command line interface. To save time and effort, we
provided users with the ability to batch submit jobs to the grid and monitor
them via the SAS® Information Delivery Portal. This method provided the
functionality with minimal work and reduced the maintenance required to
ensure that the delivered solution met the needs of the majority of our
users.
www.sasglobalforum.org/2013
21
2:00 p.m.
2:00 p.m.
Using Text Analysis to Gain Insight into Organizational
Change
Feature-Based Sentiment Analysis on Android App
Reviews Using SAS® Text Miner and SAS® Sentiment
Analysis Studio
Musthan Kader Ibrahim Meeran Mohideen, Oklahoma State
University
Jiawen Liu, Oklahoma State University
Goutam Chakraborty, Oklahoma State University
Gary Gaethe, U of Iowa
Douglas Van Daele, University of Iowa Healthcare
Paper 246-2013
Businesses often implement changes to improve customer satisfaction,
increase revenue, or improve profitability. The best situation occurs when a
business can measure the impact of the change before and after making
organizational changes. This research analyzes data from a survey of more
than 30,000 patients from a midwestern university teaching hospital. We
consider the impact of two very different changes: a move from free
parking to paid parking in 2009, and the implementation of a new online
portal designed so that patients can access their medical information. We
first analyzed the quantitative data using a key business metric and then
applied text mining and sentiment mining analysis procedures using the
qualitative data to gain deeper insights.
2:00 p.m.
So Many Films, So Little Time
Lisa Eckler, Lisa Eckler Consulting Inc.
Paper 247-2013
The Toronto International Film Festival ("TIFF") is an annual event, screening
a huge variety of new films for the international film industry as well as the
general public. The number of choices means selecting which films to order
tickets for can be overwhelming. I suffer the occupational hazard of
considering every logic problem in terms of SAS® code. Here we explore
how to use some very simple code to explore scheduling options which will
support decision-making with the goal of seeing the most films from a
priority list in the most enjoyable way. While many of us use SAS for
efficiency in our work, this is a small example of how it can also be
beneficial for personal time.
2:00 p.m.
SAS® ODS Graphics Designer - The Next Step in Amazing
Data Visualization
Christopher Battiston, Hospital For Sick Children
(Invited) Paper 248-2013
Admit it. You are swamped, overwhelmed, and desperate to find more
efficient ways of doing things. But who has the time to learn something
new? This poster won't be able to help for the majority of these issues. It
will help you become a more effective and efficient data visualization
expert, freeing up at least enough of your time to get a sandwich (and
maybe even eat it). SAS® ODS Graphics Designer is highlighted, showing
various examples with a generic step-by-step approach. Not as basic as
Graph-N-Go and not nearly as complex as SAS® Enterprise Guide®, ODS
Graphics Designer is a tool that appeals to both novice and expert users.
22
www.sasglobalforum.org/2013
Jiawen Liu, Oklahoma State University
Mantosh Kumar Sarkar, Oklahoma State University
Goutam Chakraborty, Oklahoma State University
Paper 250-2013
Sentiment analysis is a popular technique for summarizing and analyzing
consumer textual reviews about products and services. There are two major
approaches for performing sentiment analysis—the statistical model-based
approach and the Natural Language Processing NLP-based approach. In
this paper, text mining is applied first to extract the features of Android
apps. Next, the NLP approach for writing rules is used. Reviews of two
recent apps are considered; a widget app from the Brain& Puzzle category
and a game app from the Personalization category. Six hundred textual
reviews are extracted for each app from the Google Play Android App Store.
Testing results show that for both apps, the carefully designed NLP rulebased model outperforms the default statistical model for predicting
sentiments and providing deeper insights.
2:00 p.m.
Analysis of Change in Sentiments towards Chick-fil-A
after Dan Cathy’s Statement about Same-Sex Marriage
Using SAS® Text Miner and SAS® Sentiment Analysis
Studio
Swati Grover, Student
Jeffin Jacob, Student, Oklahoma State University
Goutam Chakraborty, Oklahoma State University
Paper 251-2013
Social media analysis along with text analytics is playing a very important
role in keeping a tab on consumer sentiments. Tweets posted on Twitter
are one of the best ways to analyze customers’ sentiments following any
post-corporate event. Although there are a lot of tweets, only a fraction of
them are relevant to a specific business event. This paper demonstrates
application of SAS® Text Miner and SAS® Sentiment Analysis Studio to
perform text mining and sentiment analysis on tweets written about Chickfil-A before and after the company’s president’s statement supporting
traditional marriage. We find there is a huge increase in negative
sentiments immediately following the company president’s statement. We
also track and show that the change in sentiment persists through an
extended period of time.
2:00 p.m.
Analyzing Partially Confounded Factorial Conjoint
Choice Experiments Using SAS/IML®
Song Lin Ng, Universiti Tunku Abdul Rahman
Paper 252-2013
A 2^8 partially confounded factorial design with two replicate was applied
to CCE in this study. In this study, all the responses were assumed to be
independent and hence the multinomial logit model follows. The log
likelihood is nonlinear and hence the newton-raphson method is needed to
estimate the parameters. PROC IML was used to generated the NewtonRaphson procedures. The result showed that all main effects and some of
the first-order interaction effects were significant.
2:00 p.m.
2:00 p.m.
Analyzing Partially Confounded Factorial Conjoint
Choice Experiments Using SAS/IML®
MIXED_FIT: A SAS® Macro to Assess Model Fit and
Adequacy for Two-Level Linear Models
Chin Khian Yong, Universiti Tunku Abdul Rahman
Paper 252-2013
A 2^8 partially confounded factorial design with two replicate was applied
to CCE in this study. In this study, all the responses were assumed to be
independent and hence the multinomial logit model follows. The log
likelihood is nonlinear and hence the newton-raphson method is needed to
estimate the parameters. PROC IML was used to generated the NewtonRaphson procedures. The result showed that all main effects and some of
the first-order interaction effects were significant.
2:00 p.m.
SAS Training for STD Grantees
Robert Nelson, CDC
Molly Dowling, CDC
Delicia Carey, CDC
Paper 253-2013
The efficient and effective use of STD surveillance data for programmatic
decision-making is critical to state and local STD programs. SAS software is
well-suited to help realize this goal and is available from CDC at no-cost to
state and local STD grantees. The Division of STD Prevention at CDC has
developed a SAS training course for STD grantees, SASSI, to help meet this
need. To make the training useful to users at all levels of experience, each
module is designed to stand alone. Realistic data and real-world examples
are used to help ensure relevance to the target audience and state and
local STD program staff were engaged at all phases of development. For
more information, visit http://www.cdc.gov/std/.
2:00 p.m.
A Comparison of Model Building via RPM in SAS®
Enterprise Guide® versus SAS® Enterprise Miner™
Srikar Rayabaram, Oklahoma State University
Goutam Chakraborty, Oklahoma State University
Mihaela Ene, University of South Carolina
Whitney Smiley, University of South Carolina
Bethany Bell, University of South Carolina
Paper 255-2013
When estimating multilevel models, it is important for researchers to make
sure their models fit their data. However, examining model fit can be quite
cumbersome. We have developed the macro MIXED_FIT to help researchers
assess model fit in a simple yet comprehensive way. Specifically, this paper
provides a SAS® macro that incorporates changes in model fit statistics [that
is, -2 log likelihood (-2LL), AIC, and BIC] as well as changes in pseudo-R2. By
using data from PROC MIXED ODS tables, the macro produces a
comprehensive table of changes in model fit measures and allows SAS
users to examine model fit in both nested and non-nested models, both in
terms of statistical and practical significance without having to calculate
these values by hand.
2:00 p.m.
Optimize SAS/IML® Software Codes for Big Data
Simulation
Chao Huang, Oklahoma State University
Yu Fu, Oklahoma State University
Goutam Chakraborty, Oklahoma State University
Paper 256-2013
Nowadays, real-world data volume keeps growing. Simulation also creates
large data sets. To speed up the processing of a large data set, vectorization
is a very useful code optimization skill for many matrix languages such as R,
MATLAB, and SAS/IML®. In this paper, three simulation examples in SAS/IML
are used to discuss the implementation of the latest functions and
operators from SAS/IML for vector-wise operations. The result shows that
applying vectorization in SAS/IML significantly improves the computation
performance. SAS® ODS graphics procedures are used to visualize the
results.
Paper 254-2013
2:00 p.m.
Today, most large organizations use analytics for better decision making.
Even with the widespread availability of point-and-click interfaces for
advanced predicting modeling and analytics software such as SAS®
Enterprise Miner™, building good predictive models still requires analysts to
pre-process and manipulate data. The job of pre-processing, configuring,
and comparing requires a person with deep statistical or data modeling
knowledge which large businesses can afford but this is not the case with
SMEs. SAS® contends that using RPM makes it very easy for a person with
minimal training in the area of statistics or data modeling to quickly
develop a reasonable predictive model. In this paper, we test this
contention by a controlled experiment.
Top 10 Most Powerful Functions for PROC SQL
Chao Huang, Oklahoma State University
Yu Fu, Oklahoma State University
Paper 257-2013
PROC SQL is actually not a standard SAS® procedure but a distinctive
subsystem with all features from SQL (structured query language).
Equipped with it, SAS upgrades to a full-fledging relational database
management system. In addition, PROC SQL always provides alternative
ways to manage data, besides the traditional DATA step and procedures.
SAS also supplies some goodies, such as its functions, to further strengthen
SQL operation by PROC SQL.
www.sasglobalforum.org/2013
23
2:00 p.m.
Data Set Compression Using COMPRESS=
Srinivas Reddy Busi Reddy, Oklahoma State University
Srikar Rayabaram, Oklahoma State University
Musthan Kader Ibrahim Meeran Mohideen, Oklahoma State
University
Paper 258-2013
Due to increased awareness about data mining, text mining, and big data
applications across all domains, the value of data has been realized and is
resulting in data sets with large number of variables and increased
observation size. Often it takes enormous time to process these data sets,
which can have an impact on delivery timelines. In order to handle these
constraints, think of making a large data set smaller by reducing the
number of observations, variables, or both, or by reducing the size of the
variables, without losing any of its information. In this paper, we see how a
SAS® data set can be compressed by using the COMPRESS= system option.
We also discuss some techniques to make this option more effective.
2:00 p.m.
An Integrated Approach to Codebook Generation Using
SAS®, HTML/CSS, and the .NET Framework
Helen Smith, RTI International
Mai Nguyen, RTI International
Elizabeth Eubanks, RTI International
Shane Trahan, RTI International
Paper 259-2013
For large surveys, creating comprehensive codebooks presents many
challenges. Without automation, this process becomes highly laborintensive and error-prone with data in the codebook quickly becoming
stale and failing to accurately represent underlying data sets. Another
significant challenge is that information/data for codebooks can come from
multiple sources. Such sources can include but not be limited to
questionnaire specifications, questionnaire design systems, and other
relational databases or SAS® data sets containing pertinent data. Our poster
presents an integrated approach for codebook generation using modern
tools and technologies, including SAS dictionary tables and SAS Integrated
Object Model (IOM) for data management, HTML/CSS for codebook
presentation, and the .NET framework for integrating and tying disparate
pieces together into one formatted codebook.
2:00 p.m.
Investigating the Impact of Amazon Kindle Fire HD 7” on
Amazon.com Consumers Using SAS® Text Miner and
SAS® Sentiment Analysis
Srihari Nagarajan, SAS Institute
Hari harasudhan Duraidhayalu, Kavi Associates
Goutam Chakraborty, Oklahoma State University
Paper 261-2013
This paper demonstrates the application of text mining techniques to
collect, group, and summarize positive and negative opinions of a product.
Unfortunately for popular products there are too many reviews, making it
difficult to read through all reviews and make an informed decision. For this
purpose, we developed a tool using ASP.NET to extract 1,674 customer
reviews for Kindle Fire HD 7” from Amazon.com. On the Microsoft Excel
data set thus generated, text mining can be performed to summarize
customer comments by grouping related reviews into clusters. The Text
24
www.sasglobalforum.org/2013
Parsing, Filter, Topic, and Cluster nodes are used, and outputs from every
node are discussed. Sentiment analysis is performed on the data set to
develop a model for classifying positive and negative reviews.
2:00 p.m.
Practical Application of SAS® Capabilities for Pharma
Goals and Performance Review
Ramya Purushothaman, Cognizant Technology Solutions
Paper 262-2013
This paper discusses a Pharma application that uses SAS® to leverage
internal and purchased information such as Sales and Marketing data
including drug prescriptions, dollar and unit demand, target prescribers,
and key customer account profiles to set goals, measure sales performance,
and identify trends across geography levels. The capability of SAS to handle
huge volumes of data seamlessly provides an advantage over other
technologies. The reusability of SAS macros makes SAS solutions extensible
across various brands, sales teams, and geography levels for reporting. All
of these tasks are performed through familiar Base SAS® procedures,
functions, statements, and options. The paper explains how the business
need is addressed using SAS by accessing, cleansing, and transforming
information.
2:00 p.m.
Do People Still Miss Steve Jobs As the CEO of Apple Inc.?
A Text Mining Approach: Comparing SAS® and R
Pranav Karnavat, Shanti Communication School
Anurag Srivastava, Decision Quotient
Paper 263-2013
Marketers need information on views, expressions, need and expectation of
people from social media to capitalize upon and satisfy needs and
expectation of the consumers. Twitter is a powerful social media website.
Tweets posted can be analyzed to get insights about relationships and
patterns hidden inside the textual data. In this paper tweets were collected
about Steve Jobs prior to and post his sad demise to find if customers still
miss him as the CEO of Apple Inc. using text mining technique in SAS and R.
Get tweet macro is used to fetch data from twitter in SAS while twitteR
package is to fetch data from twitter in R. To analyze data, SAS Text Miner
was used in SAS while tm package in R.
2:00 p.m.
Build Prognostic Nomograms for Risk Assessment Using
SAS®
Dongsheng Yang, Cleveland Clinic
Paper 264-2013
Nomograms from multivariable logistic models or Cox proportionalhazards regression are a popular visual plot to display the predicted
probabilities of an event for decision support. In this paper, we show how
to build a prognostic nomograms after fitting a multivariable model,
including how to assign points for each predictor under different situations
such as main effect, interaction, piecewise linear effects. Furthermore, we
also show how to use a power tool, graphic template language to construct
a nomogram Finally, a SAS® macro was developed to generate a
nomogram.
2:00 p.m.
2:00 p.m.
SAS® Enterprise Guide®: Implementation Hints and
Techniques for Insuring Success with Traditional SAS
Programmers
Life's a Song! Mining Country Music Topics Using SAS®
Text Miner
Roger Muller, Data-To-Events.Com
Paper 265-2013
Deovrat Kakde, Kavi Associates
Saurabh Ghanekar, Kavi Associates
Neetha Sindhu, Kavi Associates
There are many configuration options available in SAS® Enterprise Guide®
for both the product itself and the included advanced editor. There are also
numerous software products from SAS® that may or may not be licensed at
your site and greatly affect your workflow. Workflow options while
developing the code are numerous and range from simple line-by-line
execution up to and including the running of an entire project flow or
process. Storage of SAS code under development also deserves careful
thought. All of these topics and more are addressed to enable users to have
a very thorough non-frustrating first-time experience with SAS Enterprise
Guide. The presentation is aimed at users who have experience coding and
running SAS programs.
Paper 268-2013
2:00 p.m.
2:00 p.m.
Repairable Systems—No Longer the Stepchild of
Reliability!!! Repairable System Reliability Modeling
Using PROC RELIABILITY in SAS/QC® 9.3
Predicting Application Review Rating with SAS® Text
Miner
Deovrat Kakde, Kavi Associates
Vijitha Kaduwela, Kavi Associates
Paper 266-2013
Most assets are repairable in nature. These assets include transportation
systems such as trucks and locomotives, oil and gas drilling equipment, and
heavy engineering equipment such as earthmoving equipment. When
assets break down, they are repaired rather than replaced. The
measurement and characterization of repairable system reliability requires
a different set of statistical techniques as compared to a system that cannot
be repaired. The RELIABILITY procedure in SAS/QC® 9.2 allowed modeling
of repairable system reliability using the nonparametric mean cumulative
function (MCF). In SAS/QC 9.3, PROC RELIABILITY offers a much-needed
functionality to model recurrent event data by fitting a nonhomogeneous
Poisson process (NHPP). This paper illustrates the use of nonparametric
MCF and parametric NHPP to model reliability of critical subsystems of a
repairable system.
2:00 p.m.
Getting an Overview of SAS® Data in Three Steps
Yu Fu, Oklahoma State University
Shirmeen Virji, Oklahoma State University
Goutam Chakraborty, Oklahoma State University
Miriam McGaugh, Oklahoma State University
Paper 267-2013
For SAS programmers, one of the most important steps before
manipulating the dataset for further analysis is to get an overview of it. In
order to get an idea of the dataset, normally three areas are looked into:
variable names, statistical description, and relationship of one dataset with
other datasets within a library. The macro program introduced in this paper
writes out the names of all the variables present in a file of a particular
library, gives descriptive statistics of all the variables that are classified as
numeric, and draws a diagram to show the relationships among the
datasets. All three steps are performed by running just one program.
Rich lyrics, often with a message, are a hallmark of American country music.
Typical song topics in American country music include family, marriage,
divorce, cheating, finding love, losing love, heartbreak, happiness, drinking,
children, men, women, honky tonk, religion, politics and love of country.
This paper demonstrates the use of SAS® Text Miner to identify topics in
American country music. The lyrics of Country Music Television's (CMT) top
20 songs for the last 25 years were analyzed. The prominent topics as
identified by SAS Text Miner were compared against the tags of last.fm to
develop a measure of accuracy. The results were also validated with native
English-speaking experts.
Zhangxi Lin, The Rawls College of Business Administration,
Texas Tech University
Tianxi Dong, Rawls College of Business Administration,
Texas Tech University
Jonghyun Kim, Texas Tech University
Paper 269-2013
With the proliferation of text-based data on the Internet, there is a need for
dealing with the information overload. The large number of online user
reviews might present an obstacle to developers who want to know users'
feedback and to potential customers who are interested in applications.
Here we employ text analysis provided in SAS® Text Miner to predict the
overall and feature-based ratings for online application reviews. We use
examples from the Android Market and Apple Store, the real world of
online application stores. The findings might aid in promoting the sales of
applications by better satisfying customer demands.
2:00 p.m.
Predict the Delay in Your Airline Before They Do!
Hari harasudhan Duraidhayalu, Kavi Associates
Rajesh Inbasekaran, Kavi Associates
Paper 271-2013
This paper demonstrates the application of predictive modeling techniques
to predict the time delay in several domestic flights across the United
States. Delay in domestic flights has been a common phenomenon in the
United States and it would definitely be useful if a predictive methodology
was employed. The data set for this purpose was prepared by gathering the
past two years of data from a flight stats website. The weather details of
these airports were also collected to understand if the weather details can
be used for the prediction. By using modeling techniques such as multiple
regression, neural networks, and so on, the delay in airlines can be
predicted by knowing the airline carrier, origin, and destination airport.
www.sasglobalforum.org/2013
25
2:00 p.m.
2:00 p.m.
Calculating Subset-Weighted Analysis in PROC
SURVEYFREQ and PROC GENMOD
An Improved Data Visualization Approach for
Monitoring and Analyzing Business Performance Using
SAS/QC® Control Chart and SAS/GRAPH® Annotate
Techniques
Jessica Hale, University of Oklahoma
Paul Darden, OUHSC
David Thompson, OUHSC College of Public Health
Paper 272-2013
Stratum-specific weighted analysis is available in SAS® procedures such as
PROC SURVEYMEANS and PROC SURVEYLOGISTIC, which include the
DOMAIN statement. However, other procedures that can model correlated
outcomes, including PROC GENMOD, do not. This presentation
demonstrates a method of assigning individual weights to each record in a
data set to perform weighted subset analysis on a correlated outcome
without creating domain variables or transferring analysis to a separate
program.
2:00 p.m.
An Exploratory Graphical Method for Identifying
Associations in Sparse r x c Contingency Tables
Martin Lesser, Feinstein Institute for Medical ResearchBiostatistics
Meredith Akerman, Feinstein Institute for Medical Research
Biostatistics
Paper 273-2013
We investigate a graphical method, based on scree plots, for visualizing
“significant” departures between observed and expected cell frequencies in
RxC contingency tables, with a large number of rows and/or columns. This
method is based on Snedecor and Cochran’s (1989) proposal to identify the
cells with the largest values of (O-E)2/E, known as the contribution to chisquare. The scree plot shows the contributions plotted in descending order,
so that the user can detect which cells contribute the significant
departures, thus suggesting where the null hypothesis of independence
may have been violated. This method may be useful in large sparse RxC
tables. We used the following SAS procedures to develop a macro for
producing the scree plot: PROC FREQ (chisq, cellchi2, deviation, ODS
output), PROC SQL, and PROC GPLOT.
2:00 p.m.
SAS® Enterprise Guide®: What's in It for the Long-Term
Highly Experienced SAS® Programmer
Roger Muller, Data-To-Events.Com
Paper 274-2013
What are the benefits of SAS Enterprise Guide® as the developmental
platform for highly experienced SAS programmers who have been writing
code for a long time? This paper demonstrates a number of features that
are available in SAS Enterprise Guide for not only programming, but
viewing SAS data sets, creating multiple report outputs, improving code
storage, providing project organization and management, and more. The
techniques will emphasize the importance of the work flows in SAS
Enterprise Guide. All of these are in a state-of-the-art Microsoft Windows
environment with full copy, cut, and paste capabilities. This presentation
will focus more on the benefit to the programmer rather than on the
feature itself. The bottom line will be "What is in it for me?"
26
www.sasglobalforum.org/2013
Sheng Ding, Fedex
Baojian Guo, Fedex
Paper 275-2013
Monitoring and analyzing business performance have been proved
difficult, especially in today’s intricate business environment. Customized
Control Chart using SAS® annotate facility, however, can provide a very
useful data technique to visualize complicated business information with
manageable data visualization results. This poster introduces an improved
technique for statistical process control visualization. Combined SAS/QC®
control chart with SAS/GRAPH® annotate technique, the improved control
chart can be used to customize highlight out-of-control signals and
potential root causes. Furthermore, the authors applied customized
annotate library to leverage business impact with different the potential
root causes.
2:00 p.m.
10 SAS® Skills for Grad Student Survival: A Grad Student
“How-To” Paper
Elisa Priest, UNT HSC SPH
Paper 276-2013
Grad students learn the basics of SAS® programming in class or on their
own. Real-world research projects are usually complex and may require a
variety of different SAS tools and techniques for data exploration and
analysis. This paper is a culmination of the SAS challenges I overcame and
the SAS skills that I learned outside of the classroom. These 10 SAS skills
helped me to survive graduate school and successfully write a complex
simulation analysis in SAS for my dissertation.
2:00 p.m.
Speed it Up: Using SAS® to Automate Initial Discovery
Practices
Mariya Karimova, AdvanceMed, an NCI Company
Christine John, AdvanceMed, an NCI Company
Paper 277-2013
Healthcare investigations frequently begin with a tip containing very little
provider information. This presentation attempts to use SAS® to automate
the initial discovery process, turning a name into a full overview of the
provider. Multiple data sources are combined, which oftentimes require
fuzzy matching to resolve conflicting identifiers. The program utilizes INFILE
URL and SAS text functions to obtain meaningful information from various
websites. It further utilizes SAS ODS and SAS/GRAPH® to create a single
standard PDF report; which provides a visualization of provider billing
patterns, summarizes their affiliations, and embeds hyperlinks to original
web-based resources. Additional topics that are discussed include: creating
a script for multiple users, paramaterization, utilizing system variables, and
SAS® 9.2 to SAS® 9.3 conversion.
2:00 p.m.
2:00 p.m.
Performance Predictability By Using Social Profile in
Online P2P Lending Market
A Flexible Method to Apply Multiple Imputation Using
SAS/IML® Studio
Zhangxi Lin, The Rawls College of Business Administration,
Texas Tech University
Siming Li, Southwestern university of finance and
economics
Harshal Darade, Texas Tech University
Paper 279-2013
We study the borrower-, loan-, and group-related determinants of
performance predictability in an online P2P lending market by
conceptualizing financial and social strength to predict borrower rate and
whether the loan would be timely paid. The results of our empirical study,
conducted using a database of 9,479 completed P2P transactions in
calendar year 2007, provide support for the proposed conceptual model in
this study. The results showed that combining financial files with social
indicators can enhance the performance predictability in the P2P lending
market. Although social strength attributes do affect the borrower rate and
status, their effects are very small in comparison to the financial strength
attributes.
2:00 p.m.
Dashing out a Quick Dashboard of Graphs in SAS®
Alan Elliott, UT Southwestern
Linda Hynan, University of Texas Southwestern Medical
Center
Paper 280-2013
In a world overwhelmed with data, a challenge of a data analyst confronted
with a new data set is to produce quick and concise initial comparisons that
provide information about data distributions as well as quick statistical
comparisons on primary factors of interest. This paper combines summary
analysis graphs that incorporate statistical results in a matrix/dashboard
format on a single, concise page. SAS® users familiar with basic SAS
programming techniques will be able to produce these dashboards of
graphic results.
2:00 p.m.
Adolescent Smoking and Development of Long-Term
Habits: A Longitudinal Analysis in SAS®
Elizabeth Leslie, Kennesaw State University
Paper 281-2013
This study was an investigation into the impact of early adolescent smoking
on adult smoking habits of National Longitudinal Survey of Youth 1997
Participants over the course of 13 years. The data was from a survey
consisting of 1,212 individuals interviewed once a year for 13 years (1997 to
2009) with the frequencies and amounts of cigarettes smoked recorded.
SAS® was used for the analysis and SAS arrays, do loops and macros were
used in structuring the data. There is significant evidence that smoking
habits increase over time, sex, and age when started smoking have an
effect on number of cigarettes smoked, and the number of cigarettes
increases as the number of peers who smoke and does drugs increases.
Xue Yao, University of Manitoba
Lisa Lix, University of Manitoba
Paper 283-2013
Multiple imputation has been widely used for dealing with missing data
and measurement error problems in various scientific fields. SAS/STAT®
software offers the MI and MIANALYZE procedures for creating and
analyzing of multiple imputation data. Imputation methods in PROC MI can
be used for either continuous or classification variable with the monotone
missingness pattern and only for continuous variable with the arbitrary
missingness pattern. This paper provides an imputation method using SAS/
IML® Studio for the arbitrary missingness pattern with classification
variable. Implementing this method expands the ability to conduct
multiple imputation using SAS®.
2:00 p.m.
Growth Spline Modeling
Matthew Schuelke, Air Force Research Laboratory
Robert Terry, University of Oklahoma
Eric Day, University of Oklahoma
Paper 285-2013
In this paper we will present an extensible, hybrid statistical approach
comprised of spline modeling and growth modeling which allows for an
examination of how the relative antecedent contributions to an outcome
change through time while simultaneously controlling for past effects.
2:00 p.m.
Gee! No, GTL! Visualizing Data With The SAS Graph
Template Language
Ted Conway, Self
Paper 286-2013
When you need to produce a grid of related graphs with minimum coding,
PROC SGPANEL is hard to beat. But eventually you'll run into a situation that
demands more precise control over the output. Perhaps there are unusual
scaling/formatting requirements. Or information needs to be presented in a
specific order. Or things need to be clarified via annotations or other
markup. That's where the Graph Template Language (GTL) can help. In this
paper, we'll see how GTL can be used to create a customized grid of time
series plots from segments and measures found in the TOTARRESTS sample
data set. This may be of interest to all skill levels. It requires Base SAS, SAS
GTL, and the SAS Macro Facility on UNIX or the PC.
2:00 p.m.
Utilizing SAS® for the Construction of Preassembled
Parallel, Computerized Fixed-Test Forms under Item
Response Theory Framework
Yi-Fang Wu, Iowa Testing Programs, University of Iowa
Paper 287-2013
The preassembled, parallel computerized fixed-test (CFT) forms are among
the most popular computer-based testing models. In item response theory,
test information function plays a dominant role for designing and
comparing measurement precision of CFT forms. The current paper
develops an automated procedure by utilizing SAS® software and
procedures (i.e. PROC IML, PROC SQL, SAS/GRAPH®, GTL, and ODS) for
www.sasglobalforum.org/2013
27
constructing the CFT forms. The purpose is to demonstrate an efficient way
to obtain test and item information functions for the CFT forms and to plot
the test and item characteristic curves along with informative summary
statistics. Also, the paper investigates how measurement precision relates
to conventional item statistics. For test developers and practitioners, the
handy automated procedure through SAS and informative results are both
provided.
2:00 p.m.
2:00 p.m.
Paper 498-2013
A SAS® Macro Application for Efficient Interrupted Time
Series (ITS) Analysis Using Segmented Regression
Usefulness of text mining is now accepted worldwide to produce effective
knowledge and valuable insights of any business. Bing It On is an online
challenge offered by Microsoft allowing blind comparison of the search
results by Bing and Google. Microsoft claimed that users have chosen Bing
over Google nearly 2:1 times in these tests. Regarding this, there were
positive, negative, and mixed reactions from the vast user group, visible in
their tweets. In this research, we have collected relevant tweets using the
%GetTweet macro, and applied text mining to the data set using the SAS®
Text Miner® tab of SAS® Enterprise Miner™ 7.1 to summarize and portray the
general public opinion about this challenge and those two giant search
engines.
Sreedevi Thiyagarajan, Stanford University
Paper 288-2013
A comparison between a SAS® macro application and an existing software
tool (Joinpoint software) was conducted to identify the most efficient
software application to do a segmented regression for doing an interrupted
time series (ITS) analysis for asthma trends over time. The SAS macro
developed using the SAS 9.3 procedures NLIN and REG, when compared
with the Joinpoint software for an interrupted time series (ITS) analysis has
given an output similar to the latter and showed better running time,
efficiency as well as the time required to prepare the data sets, and total
analysis time.
2:00 p.m.
Using PROC FORMAT and Other Little Tweaks to Enable
PROC TABULATE’s Hidden Capabilities of Optimizing
Statistical Reporting
Heli Ghandehari, Baxter BioScience
Victor Lopez, Baxter Healthcare Corporation
Paper 289-2013
PROC TABULATE is arguably the most efficient approach for calculating
statistics and generating output, all within one procedure. However,
developers must often stray from PROC TABULATE when display
specifications require values to be reported as concatenated pairs. For
example, a common reporting requirement is for a mean and standard
deviation to be grouped within a single cell, with the latter enveloped by
brackets. Similarly, a range could be requested with the minimum and
maximum delimited by a dash, or perhaps a confidence interval nestled
within parentheses. The combinations are endless, but the underlying
solution is simple and universal. This paper demonstrates the utility of
PROC FORMAT’s PICTURE statement when applied in combination with
PROC TABULATE’s computational and reporting capabilities to create
customized statistical tables.
2:00 p.m.
Making it Happen: A Novel Way to Construct, Customize
and Implement Your SAS® Enterprise BI User
Enablement Framework
Tawney Moreno-Simon, Centers for Medicare & Medicaid
Services (CMS)
Vivek Seth, Computer Sciences Corporation - CSC
Paper 291-2013
Laying a solid foundation for user enablement is the holy grail of BI tool
implementation. Yet almost two-thirds (64%*) of BI Tool implementations
rate the success of user enablement initiatives “average” or lower. New BI
tool implementations struggle even further, with more than half (52%*)
rated as “fair” or “poor.”
28
www.sasglobalforum.org/2013
Application of Text Mining in Tweets to Analyze General
Opinion about “Bing It On” Challenge by Microsoft
Shreya Sadhukhan, Oklahoma State University
Taufique Ansari, Oklahoma state university
Goutam Chakraborty, Oklahoma State University
2:00 p.m.
Impact of London Olympics According to Tweeters
Yu Fu, Oklahoma State University
Shirmeen Virji, Oklahoma State University
Goutam Chakraborty, Oklahoma State University
Paper 499-2013
Overspending money on Olympics by the host country with the hope of
giving a huge boost to the economy is an age old phenomenon. The
purpose of this paper is to analyze the public sentiment on the economic
impact of London Olympics through tweets. SAS® Text Miner is employed
to summarize the collected tweets and classify them into different clusters.
Additionally, SAS® Sentiment Analysis Studio is used to corroborate our
findings and create a trend that tracks changes of public sentiments during
the London Olympics.
Quick Tips — Room 2003
2:00 p.m.
OUT= Is on the Way Out - Use ODS OUTPUT Instead
Stanley Fogleman, HARVARD CLINICAL RESEARCH
INSTITUTE
Paper 307-2013
There are, as a general rule, two methods to create a SAS data set from
procedural output. The more traditional one is the OUT= statement. This
feature is being replaced by the ODS OUTPUT statement as new capabilities
are added to procedures. In the future, only existing variables (generally in
SAS releases prior to SAS 9.1) will be available on the OUT= statement.
Therefore, it behooves the day-to-day SAS programmer to become familiar
with the new syntax.
2:15 p.m.
Implementing Metadata-driven Reports with SAS®
Stored Processes
Toby Hill, Charles Marcus Group Services
Paper 293-2013
As more organizations that use SAS® software are implementing the full
Business Intelligence reporting suite, many SAS programmers are
becoming familiar with developing SAS Stored Processes to deliver reports
for the business. Developers are often required to implement content
security in the reports or provide additional functionality for users with
specific roles. How can all this be done? One approach is to make use of the
SAS metadata. This paper demonstrates some techniques that you can
apply to your SAS code in order to make use of the SAS metadata. This will
allow you to implement security and role-based access in your SAS Stored
Process reports and minimize the amount of changes required as new users
access the platform.
function in the SAS® DATA step provides users with a simple and effective
approach to getting JSON information into SAS data sets. In this paper, two
examples of using this technique are provided.
3:15 p.m.
Back Up Your Sources During Development: A Stack of
Base SAS® Scripts
Hans Sempel, Belastingdienst (Dutch Tax and Customs
Administration)
Paper 297-2013
If you’re a Base SAS® programmer and if you ever lost your code due to
system crashes or overwriting your code, this might be the solution. The
presented code provides a means of backing up your code during
development, you can use it to save increments or you can use it for
versioning and you can restore the code you’re working on to an earlier
version.
2:30 p.m.
3:30 p.m.
Time Series Data: Anatomy of an ETL Project
Dealing with End-of-line Markers in Text Data Shared
Across Operating Systems
Leonard Polak, Wells Fargo Technology and Operations
Group
Paper 294-2013
It’s one thing to study SAS® tools and another to apply them to actual
situations. In this paper, we follow along as web data is copied and
transformed--and ultimately made available to users.
2:45 p.m.
CLISTS: Improve Efficiency of TSO Applications Using
Mainframe SAS®
Russell Hendel, Centers for Medicare and Medicaid Services
Paper 295-2013
Have you been spending a few hours every month submitting several
dozen SAS® jobs to mainframe systems using an IBM TSO environment with
the Interactive System Productive Facility (ISPF)? You know that within SAS,
SAS macros can efficiently manage repetitive tasks; but how do you
manage repetitive tasks with JCL, the TSO control language? CLIST is
precisely what you need: It enables you to automate repetitive tasks that
use JCL and SAS. CLIST is an easy language to learn, requiring no former
knowledge and using only a handful of basic commands. We present
illustrative CLIST code covering basic groups of CLIST commands. People
already familiar with JCL and SAS who write jobs using both of them will
benefit from this presentation.
3:00 p.m.
Efficient Extraction of JSON Information in SAS® Using
the SCANOVER Function
Kyong Jin Shim, Singapore Management University
Murphy Choy, Singapore Management University
Paper 296-2013
JSON or JavaScript Object Notation is a popular data interchange format
that provides a human readable format. It is language independent and can
be read easily in a variety of computer languages. With the rise of Twitter
and other types of unstructured data, there has been a move to incorporate
this data as a way of disseminating information. Twitter currently provides a
simple API for users to extract tweets using the JSON format. Although SAS
does not currently have a direct way of reading JSON, the SCANOVER
Haoyu Gu, University of Michigan
Paper 326-2013
Different operating systems use different end-of-line markers. When
sharing data across operating systems, caution must be taken. In this paper,
two examples are used to show how to read and write text data created in
Microsoft Windows from UNIX or Linux. In the examples, the use of option
TERMSTR and DLM=200Dx are discussed. The programs are run using both
SAS/CONNECT® and batch mode.
3:45 p.m.
Data Review Information: N-Levels or Cardinality Ratio
Ronald Fehd, retired
Paper 299-2013
This paper reviews the database concept: Cardinality Ratio. The SAS®
FREQUENCY procedure can produce an output data set with a list of the
values of a variable. The number of observations of that data set is called NLevels. The quotient of N-Levels divided by the number of observations of
the data is the variable's Cardinality Ratio. Its range is in (0-1]. The
Cardinality Ratio provides an important value during data review. Four
groups of values are examined.
4:00 p.m.
Healthcare Claims Processing with Base SAS® through
Denormalization of ANSI 837 Format
Victor Shigaev, CDC
Paper 300-2013
Sometimes dealing with healthcare claims can be messy. As a result of
HIPAA, all health insurance claims must be submitted to insurance payers
using the ANSI X12 837 messaging standard. This standard creates a
compact hierarchical file for quick transmission between trading partners
but because of the really complex nested structure of the data this standard
is not always easy to read in and be analyzed. The paper will give a brief
introduction to the X12 837 messaging standard, provide users a simple
way to divide raw claims data by claims through de-normalization, and a
way to use SAS® as a main tool to process and analyze the claims data.
www.sasglobalforum.org/2013
29
4:00 p.m.
Healthcare Claims Processing with Base SAS® through
Denormalization of ANSI 837 Format
Roberto Valverde, NCHS
Paper 300-2013
Sometimes dealing with healthcare claims can be messy. As a result of
HIPAA, all health insurance claims must be submitted to insurance payers
using the ANSI X12 837 messaging standard. This standard creates a
compact hierarchical file for quick transmission between trading partners
but because of the really complex nested structure of the data this standard
is not always easy to read in and be analyzed. The paper will give a brief
introduction to the X12 837 messaging standard, provide users a simple
way to divide raw claims data by claims through de-normalization, and a
way to use SAS® as a main tool to process and analyze the claims data.
4:15 p.m.
Accessing SAS® Code via Visual Basic for Applications
Jennifer Davies, Z, Inc
Paper 306-2013
SAS® software has functionality that applications such as Microsoft Access
or Excel do not have and vice versa. However, in some situations, Microsoft
applications are preferred by the user over SAS for a multitude of reasons.
This paper will discuss how to integrate the use of Microsoft applications
with the functionality of SAS programs. This becomes very important when
SAS® Business Intelligence is not available. Depending on how SAS is
installed in the user’s organization, the programmer may have to access
SAS on the PC or a server version of the application. This paper will explain
the two methods used for calling SAS code from Visual Basic for
Applications (VBA) Code (v6.5).
4:30 p.m.
Something for Nothing? Adding Group Descriptive
Statistics Using PROC SQL Subqueries
Sunil Gupta, Gupta Programming
Paper 302-2013
Can you actually get something for nothing? With PROC SQL’s subquery
and remerging features, yes, you can. When working with categorical
variables, often there is a need to add group descriptive statistics such as
group counts, minimum and maximum values for further by-group
processing. Instead of first creating the group count, minimum or
maximum values and then merging the summarized data set to the original
data set, why not take advantage of PROC SQL to complete two steps in
one? With PROC SQL’s subquery and summary functions by the group
variable, you can easily remerge the new group descriptive statistics back
to the original data set.
4:45 p.m.
FCMP -- Why?
Lisa Eckler, Lisa Eckler Consulting Inc.
Paper 298-2013
PROC FCMP allows a SAS® programmer the opportunity to create userdefined functions in SAS. Prior to the availability of FCMP in SAS 9, SAS
macros or linked routines were often used to achieve a similar –- but less
elegant -– effect. This paper examines the advantages of FCMP over the
earlier alternatives and why it is therefore so valuable to the programmer.
30
www.sasglobalforum.org/2013
Reporting and Information Visualization — Room
2002
2:00 p.m.
Virginia's Best: How to Annotate County Names and
Values on a State Map
Anastasiya Osborne, Farm Service Agency
Paper 354-2013
This paper describes a work project to annotate a Virginia state map with
long county names and National Agricultural Statistics Service (NASS) data,
using enhanced color and techniques to minimize map crowding.
Displaying text and numeric data by county on a state map is different from
displaying state-level data on a U.S. map. Long county names rather than
two-letter state abbreviations require additional effort by a programmer to
create a readable map. A SAS® program with %ANNOMAC, %CENTROID,
PROC GPROJECT, PROC GMAP, and a 20-pattern color scheme was
developed to create maps that showcased in color Virginia's top
agricultural counties. This paper is for intermediate-level programmers.
2:30 p.m.
Data Merging and Visualization to Identify Associations
Between Environmental Factors and Disease Outbreaks
Neeta Shenvi, Emory University
Xin Zhang, Emory University
Azhar Nizam, Emory University
Paper 355-2013
This paper describes data merging and visualization techniques for
epidemiological and environmental surveillance data. The ultimate goal is
to learn about associations between specific environmental factors and
disease outbreaks. In such studies, environmental and clinical surveys often
occur on different timelines. We illustrate data merging with PROC SQL to
merge environmental and clinical data with chronological lags. We use the
Graph Template Language (GTL) to demonstrate data visualizations and
correlations that enabled us to identify potential associations between
cases of the disease and environmental variables, with a variety of possible
lags.
3:00 p.m.
Introducing and Producing Thunderstorm or Rain-drop
Scatter Plots Using the SAS/GRAPH® Annotate Facility
Charlie Liu, Allergan, Inc.
Paper 357-2013
A new type of plot, the thunderstorm (or rain-drop) scatter plot is
introduced. Such a plot allows for viewing data with two or more values on
the y-axis corresponding to one value on the x-axis for each of several
subjects in a population. The resulting plot looks like rain-drops, with each
rain-drop representing data for a single subject. When data for many
subjects is plotted, it resembles a thunderstorm (hence the name). A
thunderstorm or rain-drop scatter plot is a useful tool for data visualization
and outlier detection. Using examples from clinical research, this paper
shows how to create a thunderstorm or rain-drop scatter plot by using the
SAS/GRAPH® annotate facility.
3:30 p.m.
5:00 p.m.
Do SAS® users read books? Using SAS graphics to
enhance survey research
A Concise Display of Multiple Response Items
Paper 367-2013
Surveys often contain multiple response items, such as language where a
respondent may indicate that she speaks more than one language. In this
case, an indicator variable (1=Yes, 0=No) is often created for each language
category. This paper shows how a concise tabulation of the count and
percent of respondents with a “Yes” on one or more indicator variables may
be obtained using PROC TABULATE and a MULTILABEL format. A series of
indicator variables is used to create a binary variable and its base-10
equivalent, and a MULTILABEL format is created to properly aggregate
observations with a “Yes” on two or more indicator variables. The BAND
function may also be used to easily subset observations with “Yes”
responses on certain combinations of the indicator variables.
Barbara Okerson, WellPoint
In survey research, graphics play two important but distinctly different
roles. Visualization graphics enable analysts to view respondent segments,
trends and outliers that may not be readily obvious from a simple
examination of the data. Presentation graphics are designed to quickly
illustrate key points or conclusions to a defined audience from the analysis
of the survey responses. SAS provides the tools for both these graphics
roles through SAS/Graph and ODS graphics procedures. Using a survey of
the Virginia SAS Users Group (VASUG) as the data source, this paper
answers the above question and more while illustrating several SAS
techniques for survey response visualization and presentation. The
techniques presented here include correspondence analysis, spatial
analysis, heat maps and others.
4:00 p.m.
An Innovative Approach to Integrating SAS® Macros
with GIS Software Products to Produce County-Level
Accuracy Assessments.
Audra Zakzeski, USDA NASS
Robert Seffrin, US Dept. of Agriculture
Paper 358-2013
The National Agricultural Statistics Service (NASS) produces an annual
geospatial informational data set called the Cropland Data Layer over the
U.S. detailing the land cover over each state while focusing on the vast
array of crops grown during the months of April through October. While
calculating an accuracy assessment of the land cover over an entire state is
a relatively simple process, calculating an accuracy assessment down to a
county- or crop-specific level can be extremely time-consuming. To simplify
the process, NASS created an innovative SAS® program integrating the
efficiency of the SAS Macro language with the geospatial analytical
capabilities of the GIS program ERDAS Imagine. The procedure is operated
using a SAS/AF® platform allowing analysts to easily investigate countylevel information.
4:30 p.m.
How to Become a GTL/PROC Template Jedi
Christopher Battiston, Hospital For Sick Children
Paper 359-2013
This tongue-in-cheek paper will bring together Star Wars and SAS®,
answering (at least potentially) how would SAS have been used a long time
ago in a galaxy far, far away? Using PROC TEMPLATE, GTL, and ODS,
examples will be shown of reports that could have been used by the Empire
and the Rebel Alliance. Topics will include creating reports for mobile
devices, bringing in images into the reports, and creating dynamic reports without using Jedi mind tricks on anyone!
Patrick Thornton, SRI International
Paper 360-2013
5:30 p.m.
SAS Metadata Reporting: Extracting Invaluable
Information from SAS® Metadata
Jugdish Mistry, J2L Limited
Paper 518-2013
We can see a trend in the past few releases of SAS® software; there is a big
emphasis on using and moving to using more and more metadata. It is the
one-stop place now, for all SAS applications, configuration, SAS® Data
Integration Studio, SAS® Business Intelligence, and GRID developments.
This wonderful method of storing data and managing SAS has no nice GUI
for getting this information out. So if we wanted a user list, the name of the
last person to update a SAS Data Integration flow, or list new jobs created in
the past week, we have to use the appropriate GUI and manually get this
information. This paper discusses how using SAS one could extract and
generate useful reports from metadata.
SAS Futures — Room 2018
2:00 p.m.
Your GPS for SAS® on the Cloud
Saravana Chandran, SAS
Paper 509-2013
“Cloud computing” is both mystifying and perplexing. Have you ever
wondered what cloud computing means to SAS® products and solutions?
This paper explores various possibilities of SAS products and solutions on
the cloud – public, private and hybrid – and walks through three case study
deployment models. Find out about the cloud lifecycle aspects of SAS
solutions and products.
2:30 p.m.
SAS® Virtual Applications in Your Cloud Infrastructure
Peter Villiers, SAS
Paper 510-2013
Many organizations, regardless of size, are investing in cloud
infrastructures. These infrastructures can be in-house private clouds,
commercially provided public clouds or a hybrid cloud consisting of both
types. Making purchased software run in the cloud takes time and effort to
get it right and includes its own set of challenges. If software companies
provided products as prebuilt, cloud-enabled applications, this would
benefit organizations choosing to deploy them. This paper describes an
www.sasglobalforum.org/2013
31
approach to developing prepackaged (virtual) applications for the cloud. It
outlines how these virtual applications can be deployed into various cloud
providers and integrated with existing IT resources. It also shows how these
applications can be distributed globally to provide faster communication
with users, while still being managed centrally.
3:00 p.m.
You Like What? Creating a Recommender System with
SAS®
Wayne Thompson, SAS
Paper 511-2013
Recommendation engines provide the ability to make automatic
predictions (filtering) about the interests of a user by collecting preference
information from many users (collaborating). SAS® provides a number of
techniques and algorithms for creating a recommendation system, ranging
from basic distance measures to matrix factorization and collaborative
filtering. The “wisdom of crowds” suggests that communities make better
decisions than a handful of individuals, and as a community grows, the
better its decisions are. With enough data on individual community
participation, we can make predictions about what an individual will like in
the future based on what their likes and dislikes have been in the past.
4:00 p.m.
Statistics and Data Analysis — Room 3016
10:30 a.m.
What Is Business Analytics?
J. Michael Hardin, University of Alabama
(Invited) Paper 502-2013
Analytics has become the hot, “sexy” job of the new century. The demand
for individuals with skills and expertise in the area are in great demand,
with feature articles appearing in publications ranging from The New York
Times to The Harvard Business Review to The Wall Street Journal. However,
the area has not always been so well received, especially within some
academic areas. And, even today there still remains confusion and
disagreements over the implementation and interpretation of results
obtained from the “analytic process.” This presentation will examine the
history, development, and particularly the philosophy underlying the
analytic process. Insights will be provided as to theories to understanding
and interpreting the analysis process and the associated results.
Statistics and Data Analysis — Room 2005
2:00 p.m.
Making the Most of Your SAS® Investment in the Era of
Big Data
Being Continuously Discrete (or Discretely Continuous):
Understanding Models with Continuous and Discrete
Predictors and Testing Associated Hypotheses
Paper 512-2013
(Invited) Paper 422-2013
If you follow the headlines, it doesn’t take long to recognize that big data is
a big deal. And if the amount of data that you’re faced with exceeds your
organization’s capacity for accurate and timely decision making, then it’s a
big deal for you, too! By taking advantage of massively parallel compute
environments and in-memory processing, SAS® can help you expand the
boundaries of what’s possible and transform the way you do business. This
paper helps you understand how to leverage SAS High-Performance
Analytics to explore, visualize and analyze large volumes of data.
Often a general (or generalized) linear model has both discrete predictors
(included in the CLASS statement) and continuous predictors. Binary
variables can be treated either as continuous or discrete; the resulting
models are equivalent but the interpretation of parameters differs. In many
cases, interactions between discrete and continuous variables are of
interest. This paper provides practical suggestions for building and
interpreting models with both continuous and discrete predictors. It
includes some examples of the use of the STORE statement and PROC PLM
to understand models and test hypotheses without repeating the
estimation step.
Tonya Balan, SAS
Justin Choy, SAS
5:00 p.m.
MapReduce Anywhere with DS2
Doug Sedlak, SAS
Robert Ray, SAS
Gordon Keener, SAS
Cindy Wang, SAS
Paper 398-2013
With the rise in all things Hadoop and the MapReduce programming
paradigm, you might ask, “Is the SAS® programmer left behind?” No, the
parallel syntax of the DS2 language coupled with our in-database SAS
Embedded Process and pass-through technology will allow the traditional
SAS developer to create portable MapReduce-type algorithms that are
implicitly executed in massively parallel environments, including Hadoop.
The DS2 portable syntax allows parallel algorithms to be verified on the SAS
client using sample data before releasing them on your largest problems.
32
www.sasglobalforum.org/2013
David Pasta, ICON Late Phase & Outcomes Research
Statistics and Data Analysis — Room 2007
2:00 p.m.
Using the QUANTLIFE Procedure for Survival Analysis
Guixian Lin, SAS
Bob Rodriguez, SAS
Paper 421-2013
The QUANTLIFE procedure implements quantile regression, which provides
a direct and flexible approach to modeling survival times without the
proportionality hazard constraint of the Cox model. In clinical studies,
quantile regression is helpful for identifying and distinguishing important
prognostic factors for patient subpopulations that are characterized by
short or long survival times. This paper compares the quantile regression
model with the Cox and accelerated failure time models, which are
commonly used in survival analysis. An understanding of the differences
between these approaches is essential for deciding which model to use in
practice. An example illustrates how to estimate regression parameters and
survival functions.
Statistics and Data Analysis — Room 2005
3:00 p.m.
change in price using six-month transaction-level data. Limitations and
prospects of the methods used are discussed. The inclusion of promotions
and prices of other products as covariates provides a better understanding
of the dynamics of price-demand relationships.
Computing Direct and Indirect Standardized Rates and
Risks with the STDRATE Procedure
4:00 p.m.
Yang Yuan, SAS
Paper 423-2013
In epidemiological and health care studies, a common goal is to establish
relationships between various factors and event outcomes. But outcome
measures such as rates or risks can be biased by confounding. You can
control for confounding by dividing the population into homogeneous
strata and estimating rate or risk based on a weighted average of stratumspecific rate or risk estimates. This paper reviews the concepts of
standardized rate and risk and introduces the STDRATE procedure, which is
new in SAS/STAT® 12.1. PROC STDRATE computes directly standardized
rates and risks by using Mantel-Haenszel estimates, and it computes
indirectly standardized rates and risks by using standardized morbidity/
mortality ratios (SMR). PROC STDRATE also provides stratum-specific
summary statistics, such as rate and risk estimates and confidence limits.
Statistics and Data Analysis — Room 2007
Estimating Harrell's Optimism on Predictive Indices
Using Bootstrap Samples
Yinghui Miao, NCIRE
Irena Cenzer, UCSF
Katharine Kirby, UCSF
John Boscardin, UCSF
Paper 504-2013
In aging research, it is important to develop and validate accurate
prognostic models whose predictive accuracy will not degrade when
applied in external data sources. While the most common method of
validation is split sample, alternative methods such as cross-validation and
bootstrapping have some significant advantages. The macro that we
present calculates Harrell's optimism for logistic and Cox regression models
based on either the c-statistic (for logistic) or Harrell's c (for Cox). It allows
for both stepwise and best subset variable selection methods, and for both
traditional and .632 bootstrapping methods.
3:00 p.m.
Good as New or Bad as Old? Analyzing Recurring
Failures with the RELIABILITY Procedure
Statistics and Data Analysis — Room 2007
Paper 424-2013
Creating and Customizing the Kaplan-Meier Survival
Plot in PROC LIFETEST
Bobby Gutierrez, SAS
You can encounter repeated failure events in settings ranging from repairs
of equipment under warranty to treatment of recurrent heart attacks. When
a system fails repeatedly, the risk of failure can change with each
subsequent failure—a unit once repaired or a patient once treated might
not be “good as new.” Analysis with the RELIABILITY procedure focuses
primarily on the mean cumulative function (MCF), which represents either
the average number of failures per unit over time or some related cost
measure. This paper describes how you can use PROC RELIABILITY to
estimate and compare MCFs. You can choose either a nonparametric or a
parametric approach. Features added to the procedure in SAS/QC® 12.1 are
highlighted.
Statistics and Data Analysis — Room 2005
3:30 p.m.
Price and Cross-Price Elasticity Estimation Using SAS®
4:00 p.m.
Warren Kuhfeld, SAS
Ying So, SAS
Paper 427-2013
If you are a medical, pharmaceutical, or life sciences researcher, you have
probably analyzed time-to-event data (survival data). One of several
survival analysis procedures that SAS/STAT® provides, the LIFETEST
procedure computes Kaplan-Meier estimates of the survivor functions and
compares survival curves between groups of patients. You can use the
Kaplan-Meier plot to display the number of subjects at risk, confidence
limits, equal-precision bands, Hall-Wellner bands, and homogeneity test pvalue. You can control the contents of the survival plot by specifying
procedure options with PROC LIFETEST. When the procedure options are
insufficient, you can modify the graph templates with SAS macros. This
paper provides examples of survival plot modifications using procedure
options, graph template modifications using macros, and style template
modifications.
Dawit Mulugeta, Cardinal Health
Jason Greenfield, Cardinal Health
Tison Bolen, Cardinal Health
Lisa Conley, Cardinal Health
Paper 425-2013
The relationship between price and demand (quantity) has been the
subject of extensive studies across many product categories, regions, and
stores. Elasticity estimates have also been used to improve pricing
strategies and price optimization efforts, promotions, product offers, and
various marketing programs. This presentation demonstrates how to
compute item-level price and cross-price elasticity values for two products
with and without promotions. We used the midpoint formula, the OLS
linear model, and the log-log model to measure demand response to
www.sasglobalforum.org/2013
33
Statistics and Data Analysis — Room 2005
Statistics and Data Analysis — Room 2007
4:30 p.m.
5:00 p.m.
Are You Discrete? Patients' Treatment Preferences and
the Discrete Choice Experiment
Assessing Model Adequacy in Proportional Hazards
Regression
Beeya Na, ICON Late Phase & Outcomes Research
Eric Elkin, ICON
Paper 429-2013
The discrete choice experiment (DCE) was designed for use in economics
and marketing research to study consumer preferences. DCE has been
increasingly used in health care research as a method to elicit patient
preferences for characteristics of different types of treatments. In a DCE,
attributes with varying levels are defined for treatments. Respondents are
presented with pairs of hypothetical treatments that have different
combinations of attribute levels and are asked to choose their preferred
treatment. Analyzing the responses allows evaluation of the relative
importance of the attributes and the trade-offs that respondents are willing
to make between the attributes. This paper explains how to set up the data
and discusses how to use the PHREG and LOGISTIC procedures to
appropriately analyze the conditional logit model.
Statistics and Data Analysis — Room 2007
4:30 p.m.
Cox Proportional Hazard Model Evaluation in One Shot
Polina Kukhareva, University of North Carolina at Chapel Hill
Paper 428-2013
Cox proportional hazard models are often used to analyze survival data in
clinical research. This article describes a macro that makes producing the
correct diagnostics for Cox proportional hazard models fast and easy. The
macro has three advantages over performing all the diagnostics one by
one. First, it makes it easy to run diagnostics for a long list of similar models.
Second, it allows the specification of the variables for which diagnostics
should be run. Third, it produces a comprehensive list of plots and tables
necessary for evaluation of the Cox proportional hazard model assumptions
as recommended in the SAS® course “Survival Analysis Using the
Proportional Hazards Model.” This macro can help save hours of codewriting time for a programmer who performs survival analysis.
Statistics and Data Analysis — Room 2005
5:00 p.m.
Chi-Square and t-Tests Using SAS®: Performance and
Interpretation
Jennifer Waller, Georgia Health Sciences University
Maribeth Johnson, Georgia Health Sciences University
(Invited) Paper 430-2013
Data analysis begins with data cleanup, calculation of descriptive statistics,
and the examination of variable distributions. Before more rigorous
statistical analysis begins, many statisticians perform basic inferential
statistical tests such as chi-square and t tests to assess unadjusted
associations. These tests help guide the direction of the more rigorous
analysis. This paper uses example data to show how to perform chi-square
and t tests, how to interpret the output, where to look for the association or
difference based on the hypothesis being tested, and which next steps can
be proposed for further analysis.
34
www.sasglobalforum.org/2013
Michael G. Wilson, Biostatistical Communications, Inc.
(Invited) Paper 431-2013
Proportional hazards regression has become an exceedingly popular
procedure for conducting analysis on right-censored, time-to-event data. A
powerful, numerically stable, easily generalizable model can result from
careful development of the candidate model, assessment of model
adequacy, and final validation. Model adequacy focuses on overall fitness,
validity of the linearity assumption, inclusion (or exclusion) of a correct (or
an incorrect) covariate, and identification of outlier and highly influential
observations. Due to the presence of censored data and the use of the
partial maximum likelihood function, diagnostics to assess these elements
in proportional hazards regression compared to most modeling exercises
can be slightly more complicated. In this paper, graphical and analytical
methods using a rich supply of distinctive residuals to address these model
adequacy challenges are compared.
Systems Architecture and Administration — Room
2006
10:30 a.m.
Virtualized Environment for SAS® High-Performance
Products
Tom Keefer, SAS
Rich Pletcher, SAS
Daniel Zuniga, SAS
Paper 459-2013
Virtualization technology has reached maturity, and many companies are
moving quickly to adopt these environments to support their entire IT
infrastructures. Customers are increasingly asking for reference
architectures and best practices for deploying the latest SAS® products in
virtual environments. SAS has built and tested a reference architecture that
supports a completely virtualized environment for SAS 9.3, SAS Grid
Manager and newer products such as SAS High-Performance Analytics and
SAS Visual Analytics. This paper shares results from performance testing
and best practices on how to plan, manage and deploy a successful
enterprise-class, virtualized SAS environment.
11:30 a.m.
Building a SAS® Grid Support Capability in the
Enterprise
Andy Birds, The Co-operative Banking Group
Chris Rigden, SAS
Paper 460-2013
Building SAS® Grid support capability within an organization’s IT support
function requires IT managers to consider many different aspects. There is a
need for SAS support to fit seamlessly into the Enterprise IT support model
and comply with IT policies based on standard frameworks such as ITIL,
while maintaining a level of business engagement that is far beyond that
which is required of traditional IT support teams. We will outline a practical
framework including the policies and procedures required to support a SAS
Grid in a way that provides a solid foundation that meets the immediate
and ongoing business requirements. We will discuss how to embed SAS
into an organization’s standard IT processes and how to ensure that active
business engagement is a standard activity.
3:00 p.m.
12:00 p.m.
Amy Peters, SAS
Bob Bonham, SAS
Zhiyong Li, SAS
Tips and Techniques for Deploying SAS® in an
Application Virtualization Environment
Chuck Hunley, SAS
Michael King, SAS
Casey Thompson, SAS
Rob Hamm, SAS
Paper 461-2013
Application virtualization is increasingly used by many organizations to
more easily deploy, maintain and manage their desktop applications. There
are many vendors and products on the market to choose from, including
VMWare, Citrix and Microsoft. Each vendor’s technology comes with its
unique set of features and nuances. What do you need to know to get SAS®
up and running? This paper explores some best practices, gotchas and
guidelines that will help you succeed when deploying and using SAS in an
application virtualization environment.
2:00 p.m.
Pi in the Sky: Building a Private SAS® Cloud
Andrew Macfarlane, SAS
Frank Schneider, Allianz Managed Operations and Services
SE
Paper 494-2013
In today's climate, cloud computing is a de facto term used in IT and cloud
capability a mandatory requirement for all software vendors. Based on reallife experience, this paper will discuss challenges, opportunities, and
options for developing and implementing a private SAS® cloud using SAS®
9.3. For the purposes of this paper, we focus on some essential concepts of
cloud computing including Multi Tenancy (Resource pooling); Scalability
and rapid elasticity of resources, Shared Services and the building blocks of
Platform as a Service, and suggest approaches for applying these concepts
within the SAS platform.
2:30 p.m.
Writing a Useful Groovy Program When All You Know
about Groovy Is How to Spell It
Jack Hamilton, Kaiser Foundation Hospitals
Paper 493-2013
SAS® is a powerful programming system, but it can't do everything.
Sometimes you have to go beyond what SAS provides. There are several
built-in mechanisms for doing this, and one of the newest is PROC GROOVY.
It sounds like a product of San Francisco's Haight-Ashbury, but it's actually a
programming language based on another product with San Francisco Bay
area roots, Java. You can think of it as a simplified, easier to use version of
Java -- simplified enough that you can put together a useful PROC GROOVY
program from Internet examples without knowing anything about the
language. This presentation focuses on handling directories and ZIP files,
but many other things are possible.
Monitoring 101: New Features in SAS® 9.4 for
Monitoring Your SAS® Intelligence Platform
Paper 463-2013
Ever needed an alert on SASWORK storage usage at 80 percent? Or known
when a SAS® user account has been locked out due to failed login
attempts? Or to understand the memory and swap usage of a computer
hosting the SAS Stored Process Server? What if you could see a complete
listing of process resource consumption for all physical machines hosting a
given SAS deployment? Learn about options for answering these questions,
including new tools in SAS 9.4 that autodiscover software resources
(including SAS servers and Web application servers) of your platform for
SAS Business Analytics. Discover how to use agents to collect metrics that
reflect availability, performance, utilization and throughput, giving you a
more proactive understanding of the operational state of your SAS
deployment.
3:30 p.m.
SAS® 9.3 Administration Guilty Pleasures: A Few of My
Favorite Things
Diane Hatcher, SAS
Paper 464-2013
Over the evolution of SAS® 9.3, SAS has continued to enhance and augment
its administration capabilities. Most of these capabilities are well-known
and welcome additions for SAS administrators, but there are some hidden
jewels that you may not be aware of. This paper reveals some “guilty
pleasures,” features that make the SAS environment easier to manage and
more robust, including stored process report, metadata-bound libraries,
authdomain option for libname, and special tricks with metadata folders.
4:00 p.m.
Knowing Me, Knowing My UI (ah-haa): Understanding
SAS® 9.3 and SAS® 9.4 Desktop and Web Client
Application Usage Within Your Organization
Simon Williams, SAS
Paper 465-2013
Understanding which groups of users are using which sets of desktop and
Web applications can help your organization increase efficiency while
reducing the risk to your operations. This paper details the importance of
understanding users and desktop application interactions, and the
methods by which these interactions can be defined and reported on.
Working examples using SAS® desktop and Web applications such as SAS®
Enterprise Guide®, SAS Data Integration Studio, SAS Management Console
and SAS Web Report Studio are presented.
www.sasglobalforum.org/2013
35
4:30 p.m.
SAS Visual Analytics (SASVA) and SAS High-Performance
Analytics Server (SASHPAS) - Network Considerations
and Data Management/Governance
Nicholson (Nick) Warman, Hewlett-Packard (onsite @ SAS)
(Invited) Paper 466-2013
SAS Visual Analytics (SASVA) and SAS High-Performance Analytics Server
(SASHPAS) are an entirely new approach to information analysis and
management. With these products come data network challenges/issues
and data provisioning/strategy issues. This paper begins that focused
dialogue based on over 20 years of experience with SAS and as the
engineer responsible for the world-wide configurations used by the HP(TM)
sales force, one of only two companies authorized to sell hardware to
support SASVA/SASHPAS.
5:30 p.m.
SAS® Release Management and Version Control
John Heaton, Heritage Bank
Paper 467-2013
Release Management or Application Lifecycle Management is the process
of versioning and migrating code from one environment to another in a
controlled, auditable, and repeatable process. This paper looks at the
capabilities of the current SAS® 9.3 toolset to build an effective Release
Management process within your organization.
36
www.sasglobalforum.org/2013
Applications Development — Room 2014
8:00 a.m.
Tips and Techniques for Moving SAS® Data to JMP®
Graph Builder for iPad®
Michael Hecht, SAS
Paper 008-2013
You have an iPad®. You have the JMP® Graph Builder app. But how do you
get your SAS® data set to it so you can use all the cool features that JMP
Graph Builder has to offer? This paper describes how to use the new JMP
engine in SAS to create a workflow that makes it easy to move your data
sets to the iPad.
9:00 a.m.
MACUMBA: Modern SAS® GUI Debugging Made Easy
Michael Weiss, Bayer Pharma AG
Paper 009-2013
MACUMBA is an in-house-developed application for SAS® programming. It
combines interactive development features of PC-SAS, the possibility of a
client-server environment and unique state-of-the-art features that were
always missing. This presentation covers some of the unique features that
are related to SAS code debugging. At the beginning, special code
execution modes are discussed. Afterwards, an overview of the graphical
implementation of the single-step debugger for SAS macros and DATA step
is provided. Additionally, the main pitfalls of development are discussed.
9:30 a.m.
Give the Power of SAS® to Excel Users Without Making
Them Write SAS Code
William Benjamin, Owl Computer Consultancy LLC
Paper 010-2013
Merging the ability to use SAS® and Microsoft Excel can be challenging.
However, with the advent of SAS® Enterprise Guide®, SAS® Integration
Technologies, SAS® BI Server software, JMP® software, and SAS® Add-In for
Microsoft Office; this process is less cumbersome. Excel has the advantages
of being cheap, available, easy to learn, and flexible. On the surface, SAS
and Excel seem widely separated without these additional SAS products.
But wait, BOTH SAS AND EXCEL CAN INTERFACE WITH THE OPERATING
SYSTEM. SAS can run Excel using the command and Excel can run SAS as an
“APPLICATION.” This is NOT DDE; each system works independently of the
other. This paper gives an example of Excel controlling a SAS process and
returning data to Excel.
10:00 a.m.
Automated Testing of Your SAS® Code and Collation of
Results (Using Hash Tables)
Andrew Ratcliffe, RTSL.eu
Paper 011-2013
Testing is an undeniably important part of the development process, but its
multiple phases and approaches can be under-valued. I describe some of
the principles I apply to the testing phases of my projects and then show
some useful macros that I have developed to aid the re-use of tests and to
collate their results automatically. Tests should be used time and again for
regression testing. The collation of the results hinges on the use of hash
tables, and the paper gives detail on the coding techniques employed. The
small macro suite can be used for testing of SAS® code written in a variety
of tools including SAS® Enterprise Guide®, SAS® Data Integration Studio,
and the traditional SAS Display Manager Environment.
10:30 a.m.
A Metadata-Driven Programming Technique Using SAS®
Xiyun (Cheryl) Wang, Statistics Canada
Paper 012-2013
In a typical SAS® system, validations on user inputs and setting defaults for
missing values in inputs are essential to ensure that the system can
continue its processing without errors. This paper describes, in a SAS
system, how to define validation rules as metadata for various type of
inputs such as library, data sets, etc., and how to register default values for
missing values for inputs as metadata. Furthermore, it illustrates how to use
those metadata to automate the validation processes and data imputation
process. It provides a SAS programming technique to ease system
development with efficiency, re-usability, easy maintainability, and coding
consistency.
11:00 a.m.
Knowing When to Start, Where You Are, and How Far
You Need to Go: Customized Software Tracks Project
Workflow, Deliverables, and Communication
Eric Vandervort, Rho
Paper 013-2013
In a clinical trials environment, projects can have multiple statisticians and
statistical programmers working on tables, listings, and figures, or
"displays", for project deliverables. Communication between the various
team members regarding when to program, validate, review, or revise
these displays is vital to the success of a project. This paper describes a
custom web-based application that stores relevant data about displays,
tracks programming and reviewing workflow, and provides a tool for
project-level management overview.
11:30 a.m.
Extension Node to the Rescue of the Curse of
Dimensionality via Weight of Evidence (WOE) Recoding
Satish Garla, SAS
Goutam Chakraborty, Oklahoma State University
Andrew Cathie, SAS
Paper 014-2013
Predictive models in data mining applications often involve very large data
sets that contain numerous categorical variables with large numbers of
levels. These models often suffer from the curse of dimensionality.
Enhanced weight of evidence (WOE) methods can be used to effectively
incorporate high-dimensional categorical inputs into a data mining model.
Weight of evidence technique converts a nominal input into an interval
input by using a function of the distribution of a target variable for each
level in the nominal input variable. SAS® Enterprise Miner™ has a facility to
create extension nodes that work in the same way as a usual node. This
paper explains creation of an extension node in SAS Enterprise Miner that
performs WOE recoding.
www.sasglobalforum.org/2013
37
1:30 p.m.
4:30 p.m.
Take Home the ODS Crown Jewels: Master the New
Production Features of ODS LAYOUT and Report Writing
Interface Techniques
Extraction, Transformation, and Loading (ETL) for
Outcome Measures of Workers’ Compensation Benefits
Paper 015-2013
Paper 018-2013
The ODS “crown jewels” are the most sought-after features that allow the
customer to create sophisticated documents that can readily be published
via print, Web or mobile device. Journey with the SAS® 9.4 release as we
explore the new enhancements in this production release of ODS LAYOUT
and the Report Writing Interface. This example-driven content is intended
to empower and captivate the novice ODS customer while challenging
even the most advanced user.
Base SAS® was used to create a data sub-system for measuring outcomes,
added to a data system (coded in SAS) of benefit costs and employment.
One claim per injured worker per fiscal year is extracted as a study or
control record, using business-rule code. Disability benefits and
employment data are transformed to time-series records for claims, which
are transformed to time-series statistics by fiscal year. Programs are run
remotely on a UNIX data warehouse, and SAS data sets and metadata are
loaded to the warehouse and downloaded to a LAN. Quarterly generations
are kept for analysis of claim development.
Dan O'Connor, SAS
2:30 p.m.
Set Yourself Free: Use ODS Report Writing Technology in
SAS® Enterprise Guide® Instead of Dynamic Data
Exchange in PC SAS®, Part II SAS Code Revealed
Robert Springborn, Office of Statewide Health Planning &
Development
(Invited) Paper 016-2013
The ability to prepare custom designed reports and convey your message
in a clear and concise manner is very important in today’s sophisticated
business environment. Traditional use of Dynamic Data Exchange (DDE) in
PC SAS® to produce custom designed reports is the result of widespread
and popular use of Microsoft Excel. However with most business
organizations transitioning to SAS® Enterprise Business Intelligence (EBI),
where DDE is not compatible, ODS Report Writing technology is a powerful
alternative to create custom designed reports in SAS® Enterprise Guide®.
The driving force for this topic was the need to create hospital-level data
discrepancy reports which compare clinical data to administrative data to
verify risk factors used in a risk-adjusted operative mortality model
3:30 p.m.
Mike Maier, Oregon Department of Consumer and Business
Services
5:00 p.m.
Predictive Modeling in Sports Leagues: An Application in
the Indian Premier League
Pankush Kalgotra, Oklahoma State University
Ramesh Sharda, Oklahoma State University
Goutam Chakraborty, Oklahoma State University
Paper 019-2013
The purpose of this article is to develop models that can help team
selectors build talented teams with minimum possible spending. In this
study, we build several predictive models for predicting the selection of a
player in the Indian Premier League, a cricket league, based on each
player’s past performance. The models are developed using SAS® Enterprise
Miner™ 7.1. The best-performing model in the study is selected based on
the validation data misclassification rate. The selected model provides us
with the probability measure of the selection of each player, which can be
used as a valuation factor in the bidding equation. The models that are
developed can help decision-makers during auction set salaries for the
players.
The SAS® Output Delivery System: Boldly Take Your Web
Pages Where They Have Never Gone Before!
Beyond the Basics — Room 2016
Paper 017-2013
8:00 a.m.
Chevell Parker, SAS
HTML output is one of the most effective and portable methods of
disseminating information within an organization. Using a variety of
techniques in the SAS® Output Delivery System (ODS), you can create HTML
output that increases the visibility and functionality of your web pages. This
paper discusses how to use those ODS techniques to deliver web content
specifically for mobile devices, web content for both mobile and desktop
devices, and web content specifically for desktop devices. In the first two
categories, the paper discusses the challenges and solutions for both types
of web content. For desktop devices, the paper discusses how to extract
data from web pages and place it into pivot tables for data-visualization
purposes in Microsoft Excel.
38
www.sasglobalforum.org/2013
Be Prompt: Do it Now! Creating and Using Prompts in
SAS® Enterprise Guide®
Ben Cochran, The Bedford Group
(Invited) Paper 028-2013
Prompts are a quick and powerful way to give your programs, tasks, and
projects in SAS® Enterprise Guide® interactive capabilities. By putting
prompts in your code, you increase your ability to reuse code and also
enable the code to be customized using the value that is entered through
the prompt. Prompts are fairly easy to create, and this paper takes a stepby-step approach that explains how to create and use prompts.
9:00 a.m.
2:00 p.m.
Have it Your Way: Creating Reports with the DATA Step
Report Writing Interface
RUN_MACRO Run! With PROC FCMP and the
RUN_MACRO Function from SAS® 9.2, Your SAS®
Programs Are All Grown Up
Pete Lund, Looking Glass Analytics
(Invited) Paper 040-2013
SAS provides powerful, flexible reporting procedures. ODS provides
enormous control over the appearance of procedure output. Still, for times
where you need more, the Report Writing Interface can help. “Report
Writing Interface” simply refers to using the ODSOUT object in a DATA step.
This allows you to lay out the page, create tables, embed images, add titles,
and more using any desired DATA step logic. Most style capabilities of ODS
are available, so your output can have fonts, colors, backgrounds, and
borders to customize your report. This presentation will cover the basics of
the ODSOUT object and then walk through techniques to create four “real
world” examples. You might even go home and replace some PROC
REPORT code!
10:00 a.m.
Inventory Your Files Using SAS®
Brian Varney, Experis Business Analytics
(Invited) Paper 030-2013
Whether you are attempting to figure out what you have when preparing
for a migration or you just want to find out which files or directories are
taking up all of your space, SAS® is a great tool to inventory and report on
the files on your desktop or server. This paper presents SAS code to
inventory and report on the location you want to inventory.
11:00 a.m.
This Is the Modern World: Simple, Overlooked SAS®
Enhancements
Bruce Gilsen, Federal Reserve Board
(Invited) Paper 031-2013
Some smaller, less dramatic SAS® enhancements seem to fall through the
cracks. Users continue to employ older, more cumbersome methods when
simpler solutions are available. This includes enhancements introduced in
SAS 9.2, SAS 9, SAS 8, or even SAS 6! This paper reviews underutilized
enhancements that allow you to more easily 1. Write date values in the
form yyyymmdd. 2. Increment date values with the INTNX function. 3.
Create transport files: PROC CPORT/CIMPORT versus PROC COPY with the
XPORT engine. 4. Count the number of words or the number of occurrences
of a character or substring in a character string. 5. Concatenate character
strings. 6. Check if any of a list of variables contains a value. 7. Sort by the
numeric portion of character values. 8. Retrieve DB2 data on z/OS
mainframes.
1:30 p.m.
Submitting SAS® Code on the Side
Rick Langston, SAS
Paper 032-2013
Dylan Ellis, Mathematica Policy Research
Paper 033-2013
When SAS® first came into our life, it comprised but a DATA step and a few
procedures. Then we trained our fledgling programs using %MACRO and
%MEND statements, and they were able to follow scripted instructions. But
with SAS 9.2 and 9.3, your macros are now wearing the clothes of a PROC
FCMP function; you no longer need to feed every parameter with a spoon.
These functions are independent programming units, and this talk shows
how they can be put to use for handy calculations, standardizing and
simplifying code, and adding dynamic new capabilities that may change
the way you program.
2:30 p.m.
A Flock of C-Stats, or Efficiently Computing Multiple
Statistics for Hundreds of Variables
Steven Raimi, Magnify Analytic Solutions
Bruce Lund, Marketing Associates, LLC
Paper 034-2013
In other presentations, the authors have provided macros that efficiently
compute univariate statistics for hundreds of variables at a time. The classic
example is when a modeler must fit a binary model (two-valued target) and
has available hundreds of potential numeric predictors. Such situations may
occur when third-party data sets are added to in-house transactional data
for direct marketing or credit scoring applications. The paper describes the
SAS® code to compute these statistics, focusing on the techniques that
make these macros efficient. Topics include macro techniques for
identifying and managing the input variables, restructuring the incoming
data, and using hash objects to quickly count the number of distinct values
for each variable.
3:00 p.m.
A Better Way to Flip (Transpose) a SAS® Data Set
Arthur Tabachneck, myQNA, Inc.
Xia Keshan, Chinese Financial electrical company
Robert Virgile, Robert Virgile Associates, Inc.
Joe Whitehurst, High Impact Technologies
Paper 538-2013
Many SAS® programmers have flipped out when confronted with having to
flip (transpose) a SAS data set, especially if they had to transpose multiple
variables, needed transposed variables to be in a specific order, had a
mixture of character and numeric variables to transpose, or if they needed
to retain a number of non-transposed variables. Wouldn’t it be nice to have
a way to accomplish such tasks that was easier to understand and modify
than PROC TRANSPOSE, was less system resource-intensive, required fewer
steps, and could accomplish the task as much as fifty times or more faster?
This paper explains the new DOSUBL function and how it can submit SAS®
code to run “on the side” while your DATA step is still running. It also
explains how this function differs from invoking CALL EXECUTE or invoking
the RUN_COMPILE function of FCMP. Several examples are shown that
introduce new ways of writing SAS code.
www.sasglobalforum.org/2013
39
3:30 p.m.
9:00 a.m.
Big Data, Fast Processing Speeds
Data Entry in SAS® Strategy Management: A New, Better
User (and Manager) Experience
Kevin McGowan, SAS
Paper 036-2013
David Shubert, SAS
As data sets continue to grow, it is important for programs to be written
very efficiently to make sure no time is wasted processing data. This paper
covers various techniques to speed up data processing time for very large
data sets or databases, including PROC SQL, data step, indexes and SAS®
macros. Some of these procedures may result in just a slight speed increase,
but when you process 500 million records per day, even a 10 percent
increase is very good. The paper includes actual time comparisons to
demonstrate the speed increases using the new techniques.
Paper 052-2013
Beyond the Basics — Room 3016
9:30 a.m.
4:30 p.m.
Versatile Global Prompting for SAS® Web Report Studio
Hong Jiang, Deloitte
Maximizing the Power of Hash Tables
David Corliss, Magnify Analytic Solutions
(Invited) Paper 037-2013
Hash tables provide a powerful methodology for leveraging bid data by
formatting an n-dimensional array with a single, simple key. This
advancement has empowered SAS® programmers to compile exponentially
more missing data points than ever before, creating tables with hundreds
of fields of all types in which the majority of data in this vast array is empty.
However, the hash structure also supports analytics to calculate maximum
likelihood estimates for missing values, leveraging extensive data resources
available for each individual. An important application of this is in
sentiment analysis, where social media text expresses likes or dislikes for
particular products. Customer data, including sentiments for other
products, are used to model sentiment where an individual’s preference
has not been made known.
Business Intelligence Applications — Room 2009
8:00 a.m.
SAS® BI Dashboard: Interactive, Data-Driven Dashboard
Applications Made Easy
Scott Sams, SAS
Paper 061-2013
The latest release of SAS® BI Dashboard gives you powerful new
functionality for designing dashboard applications, which are a set of
interactive dashboards that present a data-driven story to the user.
Previously, dashboards have been able to link to other applications and
pass parameters. Now with version 4.31 M2, dashboards can link to other
dashboards and pass the current click context as parameters, guiding your
users through their data with customized dashboard presentations. The
context parameters can even be passed to a stored process and have its
generated data presented by a custom dashboard. This paper shows you
how to build a simple dashboard application using key enhancements in
SAS BI Dashboard 4.31 M2.
40
www.sasglobalforum.org/2013
Data entry in SAS® Strategy Management has never been an especially
pleasant task given the outdated user interface, lack of data validation and
limited workflow management. However, release 5.4 unveils a complete
overhaul of this system. The sleeker HTML5-based appearance provides a
more modern Web experience. You are now able to create custom data
validation rules and attach a rule to each data value. Discover how form
workflow is supported via SAS Workflow Studio and the SAS Workflow
Services.
Paper 054-2013
Prompts built into the information map are convenient tools for developers
using SAS® Web Report Studio. However, the parameter values set through
these prompts are not able to populate to other web report sections or
objects, limiting their usefulness. This paper describes a solution for
creating versatile global prompts that support one-time user response for
multiple objects or sections in a SAS web report. Both single-value and
multiple-value selection features can be implemented by following the
directions described in this paper.
10:00 a.m.
Big Data - Dream IT. Build IT. Realize IT.
Paul Kent, SAS
Andy Mendelsohn, Oracle Corporation
Maureen Chew, Oracle Corporation
Steven Holmes, Bureau of Labor Statistics
(Invited) Paper 488-2013
This session will present unique perspectives on building solutions for Big
Data architectures to enable turning the vision into reality. Guest speakers
Andrew Mendelsohn, Senior Vice President, Oracle Database Server
Technologies, and Paul Kent, SAS Vice President, Big Data, will discuss
collaborative efforts towards best-of-breed Big Data analytic solutions and
convergence of game-changing IT strategies. We'll also hear from the
Bureau of Labor & Statistics on how their SAS® usage produces one of the
mostly watched economic series (U.S. employment) each month.
11:00 a.m.
How to List All Users That Have Access to a SAS®
Information Delivery Portal 4.3 Page
Bernt Dingstad, If Insurance
Paper 055-2013
This paper describes how to access SAS® Metadata from a Base SAS® client
and make simple listings of often very urgent information and in the end
distribute this information utilizing the SAS Enterprise BI framework.
11:30 a.m.
Key Aspects to Implement A Perfect SAS® BI Platform
Interact between objects. - Drill and expand. - Collaborate. - Share results
on the Web and mobile devices. SAS Visual Analytics provides fast, effective
dashboards anywhere, regardless of how big your data may be.
Paper 489-2013
4:30 p.m.
A perfect SAS® architecture is not defined just by successful installation of a
SAS platform but also by ensuring good performance, easy maintenance,
compliance to all security, secured environment, scalability, good
administration practices, proper monitoring, seamless Integration with
Interfacing system, etc. SAS provides lot of flexibility in order to integrate
with other interfacing systems; however, a perfect SAS Enterprise
Implementation is not just driven with the maturity of SAS platform but
also requires matured implementation of other interfacing platforms. SAS
user experience starts from the first click on the SAS client and driven from
SAS environment capabilities and its integration to interfacing systems.
Hence interfacing systems also have a key role to play in order to get
perfect SAS architecture.
How to Automate Security Filters for SAS® OLAP Cubes
Using Users Groups Information Available in SAS®
Management Console
Gaurav Agrawal, Major Financial Company
1:30 p.m.
Whirlwind Tour Around SAS® Visual Analytics
Anand Chitale, SAS
Christopher Redpath, SAS
Paper 057-2013
SAS brings a revolutionary approach for analytical- based data visualization,
exploration and reporting. Come and join us for a whirlwind tour through
SAS® Visual Analytics 6.1, inspired by customer experiences and use cases.
We start with an overview of the technology concept behind SAS Visual
Analytics and enter into the Hub, leading to various destinations, including
the Data Builder, Explorer and Designer, ending the journey with Mobile
integration. We also offer a sneak peek into world of Microsoft Office. Don’t
worry; SAS coders are not left behind. The final stop is SAS® Enterprise
Guide®, where you learn how to leverage the technology behind SAS Visual
Analytics through SAS code.
2:30 p.m.
The Forest and the Trees: See It All with SAS® Visual
Analytics Explorer
Nascif Abousalh-Neto, SAS
Plinio Faria, Bradesco
Paper 048-2013
In order to limit the data that a user can access in an OLAP Cube, it is
required to use MDX conditions, and those expressions must be
customized for every OLAP Cube because each one has a different
structure. This paper will focus on showing how to automate the creation of
MDX conditions using the users groups information available in the
metadata server and in a way that is possible to change the data subset
that the users are allowed to see only changing the user group in SAS®
Management Console. It will be demonstrated also how to implement
OLAP Cube security adding directly the users login to a cube dimension.
5:00 p.m.
From Factbook to Dashboard in T Minus No Time and
Counting!
Alicia Betsinger, Office of Strategic Initiatives, UT System
Annette Royal, The University of Texas System
Paper 049-2013
The University of Texas System has been publishing detailed data on
institutional performance for years using static PDF files and Excel
documents. With requests for more data increasing, this approach was
unsustainable. The Office of Strategic Initiatives (OSI) was spending too
much time collecting and processing data for the Chancellor, Board of
Regents, and media. There was no time for in-depth research or analysis.
Instead of users using the data to help support better management, the
data was managing the users. What grew from an internal office’s need
morphed into a larger UT System need for a BI system that would support
the Chancellor’s framework by providing an accessible and customizable
tool for monitoring institutional performance and progress toward
transparency and accountability goals.
Paper 058-2013
It’s a jungle out there! A data jungle, that is. With so much data to process,
it’s too easy to get lost. What’s a data explorer to do? This paper explains
how data exploration journeys usually follow a generic workflow composed
of seven well-defined tasks that are easy to perform using SAS® Visual
Analytics Explorer. Armed with this knowledge, you will be able to see both
the forest and the trees, and never worry again about losing your way.
3:30 p.m.
Fast Dashboards Anywhere with SAS® Visual Analytics
Rick Styll, SAS
Paper 059-2013
Dashboards can be the best starting point to provide a high-level view of
the most relevant information to monitor, analyze and collaborate around
business performance. A growing number of organizations are creating
dashboards to improve fact-based decision making, but if the dashboards
fail to return results quickly or are difficult to understand, user adoption will
soon wane. SAS® Visual Analytics includes an authoring interface called
Designer that lets you build responsive, well-formatted and effective
dashboards that allow users to: - Place, size and style objects precisely. -
Data Management — Room 2001
8:00 a.m.
Data Fitness: A SAS® Macro-based Application for Data
Quality of Large Health Administrative Data
Mahmoud Azimaee, Institute for Clinical Evaluative Sciences
(Invited) Paper 075-2013
This paper introduces a SAS® macro-based application package as a
solution for creating automated data quality assurance reports for large
health administrative data. It includes methods and tools for developing
metadata for a SAS data holding, for measuring different data quality
indicators using a Data Quality Framework, and for generating automated
visual data quality reports. Because quality of data documentation should
be considered as a usability and interpretability factor for good quality data,
this application uses the same metadata developed for data quality
purposes to generate an automated web-based data dictionary as well.
www.sasglobalforum.org/2013
41
9:00 a.m.
Adaptive In-Memory Analytics with Teradata: Extending
Your Existing Data Warehouse to Enable SAS® HighPerformance Analytics
Greg Otto, Teradata Corporation
Tom Weber, SAS
Paper 076-2013
SAS® High-Performance Analytics rapidly analyzes big data in-memory. The
Initial High-Performance Analytics SAS offering on Teradata co-locates SAS®
on the database nodes in a separate appliance. Data is replicated to the
appliance for use by the SAS analytics. SAS and Teradata have developed a
new in-memory analytics architecture that delivers the power of SAS HighPerformance Analytics to data in the Enterprise Data Warehouse, without
replication to an appliance. In this “Asymmetric” architecture, dedicated
SAS nodes access Teradata data on demand via high-speed networking.
Asymmetric in-memory processing extends a Teradata EDW to support SAS
High-Performance Analytics with minimal impact. This paper explains the
Asymmetric architecture and configuration options, and quantifies the
impact to Teradata systems that are extended to support SAS in-memory
analytics.
needed. This paper explains how SAS® Data Quality functions can be
invoked in database, eliminating the need to move data, thus delivering
data quality that meets the need for near-real-time performance for today’s
business. Graphic results comparing performance metrics of traditional
data quality operations against in-database data quality will be presented
along with details on how these results scale with database resources.
1:30 p.m.
The SQL Tuning Checklist: Making Slow Database
Queries a Thing of the Past
Tatyana Petrova, SAS
Jeff Bailey, SAS
Paper 080-2013
The DB2 SQL query had been running for 36 hours before it was killed.
Something had to be done, but what? Chances are, you’ve been there and
felt the pain and helplessness. Is there anything you can do to make it run
faster? Absolutely. This presentation teaches the steps and the mindset
required to solve this all-too-common problem. We cover specific
techniques that will enable you to solve the problem and discuss DB2,
Oracle, Greenplum and Teradata.
9:30 a.m.
2:30 p.m.
Best Practices in SAS® Data Management for Big Data
Need for Speed - Boost Performance in Data Processing
with SAS/Access® Interface to Oracle
Malcolm Alexander, SAS
Nancy Rausch, SAS
Paper 077-2013
This paper discusses capabilities and techniques for optimizing SAS® data
management products for big data. It demonstrates how SAS supports
emerging Apache Hadoop technologies, including details on the various
languages available in the platform and how to use them with SAS. An
overview of SAS Data Management integration with the SAS® LASR™
platform is also provided. Practical tips and techniques are presented,
including case studies that illustrate how to best use the SAS data
management feature set with big data.
10:30 a.m.
SAS® Data Integration Studio: The 30-Day Plan
Svein Erik Vralstad, Knowit Desicion Oslo AS
Paper 081-2013
Big Data is engulfing us. The expectations of users increase, and analytics is
getting more and more advanced. Timely data and fast results have never
had greater value. In data warehousing, analytics, data integration, and
reporting, there is an ever-growing need for speed. When operating in
environments where performance is of importance, it is of great value to
fully understand the interaction between the different components of the
environment. Hence, the importance of in-database execution is
accelerating. To know when to let SAS process data, and when to use
Oracle to perform the task is then of great value. This paper explores ways
to achieve substantial gains in performance when processing (read,
transform, calculate, and write) data in an effective manner.
John Heaton, Heritage Bank
3:00 p.m.
Paper 078-2013
Sharpening Your Skills in Reshaping Data: PROC
TRANSPOSE vs. Array Processing
When starting your journey with SAS® Data Integration Studio, it is
important to get the basics correct. This paper outlines the main framework
and activities needed within the first 30 days to set you up for success using
SAS Data Integration Studio.
11:00 a.m.
In-Database Data Quality: Performance for Big Data
Scott Gidley, SAS
Mike Frost, SAS
Charlotte Crain, SAS
Paper 079-2013
Data quality and high performance have joined forces. Today is an era of
big data, extremely large data warehouses and potential security issues for
moving data. Traditional data quality is performed with an ETL-like
operation of extracting, processing and publishing back to the source. To
guarantee high performance and assure data security, a new approach is
42
www.sasglobalforum.org/2013
Arthur Li, City of Hope
(Invited) Paper 082-2013
A common data managing task for SAS® programmers is transposing data.
One of the reasons for performing data transformation is that different
statistical procedures require different data shapes. In SAS, two commonly
used methods for transposing data are using either the TRANPOSE
procedure or array processing in the DATA step. Using PROC TRANSPOSE
mainly requires grasping the syntax and recognizing how to apply different
statements and options in PROC TRANSPOSE to different types of data
transposition. On the other hand, utilizing array processing in the DATA
step requires programmers to understand how the DATA step processes
data during the DATA step execution. In this talk, these two methods will be
reviewed and compared through various examples.
4:00 p.m.
8:30 a.m.
Pulling Data from the Banner Operational Data Store
with SAS® Enterprise Guide: Not Only Fast but Fun!
Targeting Public Value in New Zealand
Michael O'Neil, Ministry of Social Development
Paper 083-2013
The New Zealand Ministry of Social Development is implementing what has
been called the “Investment-Based Approach,” which aims to improve
social sector performance through better targeting. Social and fiscal
outcomes can be better achieved through smart targeting using clientcentred evidence to inform strategic and case-level targeted decisions. This
paper describes progress to date.
Claudia McCann, East Carolina University College of Nursing
The assessment of learning and of services in higher education is crucial for
continued improvement. Administrators and faculty, therefore, require data
for their decision-making processes. There are many data input experts on
campus and, unfortunately, far fewer who can extract the data in the
aggregate form required by administrators, accreditors, and other
institutional stakeholders. The SAS® Enterprise Guide interface with the
Banner Operational Data Store is a very powerful combination of softwares
that enable the end user to quickly access the institution's data and
produce reports. More powerful still is the ability to bring other relational
databases, such as Excel spreadsheets, into the SAS Enterprise Guide
environment, thereby allowing variables not available in the Operational
Data Store to be used in comparative analyses.
4:30 p.m.
Best Practices in Enterprise Data Governance
Nancy Rausch, SAS
Scott Gidley, SAS
Paper 084-2013
Data governance combines the disciplines of data quality, data
management, data policy management, business process management
and risk management into a methodology that ensures important data
assets are formally managed throughout an enterprise. SAS has developed
a cohesive suite of technologies that can be used to implement efficient
and effective data governance initiatives, thereby improving an enterprise’s
overall data management efficiency. This paper discusses data governance
best practices. It explains where and how SAS® capabilities (such as the
business data network, reference data management, data federation, data
quality, data management and master data management) can be used to
ensure data governance initiatives remain successful, continue to deliver
overall return on investment, and gain buy-in across the enterprise.
Data Mining and Text Analytics — Room 2004
8:00 a.m.
Information Retrieval in SAS®: The Power of Combining
Perl Regular Expressions and Hash Objects
Lingxiao Qi, Kaiser Permanente
Fagen Xie, Kaiser Permanente
Paper 091-2013
The volume of unstructured data is rapidly growing. Effectively extracting
information from huge amounts of unstructured data is a challenging task.
With the introduction of Perl regular expressions and hash objects in SAS®
9, the combination of these two features can be very powerful in
information retrieval. Perl regular expressions can be used to match and
manipulate various complex string patterns. The hash object provides an
efficient and convenient mechanism for quick data storage and retrieval. By
leveraging the best from both tools and applying it to the electronic
medical data, we show how pattern searching on free text is made easy
while reducing coding effort and increasing performance.
Paper 092-2013
9:00 a.m.
Using the Boosting Technique to Improve the Predictive
Power of a Credit Risk Model
Andres Gonzalez, Colpatria / Scotia Bank
Darwin Amezquita, Colpatria-Scotia Bank
Alejandro Correa Bahnsen, Luxembourg University
(Invited) Paper 093-2013
In developing a predictive model, the complexity of the population used to
build the model can lead to very weak scorecards when a traditional
technique such as logistic regression or an MLP neural network is used. For
these cases some nontraditional methodologies like boosting could help
improve the predictive power of any learning algorithm. The idea behind
this technique is to combine several weak classifiers to produce a much
more powerful model. In this paper, boosting methodology is used to
enhance the development of a credit risk scorecard in combination with
several different techniques, such as logistic regression, MLP neural
networks, and others, in order to compare the results of all methodologies
and determine in which cases the boosting algorithm increases model
performance.
9:30 a.m.
Using the Power of SAS® to Analyze and Solve Quality
Problems at Shanghai General Motors
Yu Zhang, Shanghai General Motors Co.,Ltd
Ying Wang, Shanghai General Motors Co.,Ltd
Shaozong Jiang, Information System Department,Shanghai
General Motors Co.,Ltd
Yahua Li, Information System Department,Shanghai General
Motors Co.,Ltd
Jiawen Zhang, Information System Department, Shanghai
General Motors Co.,Ltd
Nanxiang Gao, Quality Departmentï¼Shanghai General
Motors Co.,Ltd
Jian Li, Quality Departmentï¼Shanghai General Motors
Co.,Ltd
Minghua Pan, Quality Departmentï¼Shanghai General
Motors Co.,Ltd
Paper 090-2013
Data to assist in solving quality problems is of enormous value to quality
departments in the automotive industry, including that of Shanghai
General Motors (SGM). However, millions of claims records, tens of
thousands of solving reports, dozens of language descriptions, and
heterogeneous regional code present great difficulty for dataflow and
knowledge management. This paper explores SGM’s information system,
www.sasglobalforum.org/2013
43
known as Problem Solving Analysis (PSA), which uses several foundation
tools of SAS®, such as Base SAS®, SAS/CONNECT®, and SAS/ACCESS®, to
solve business problems faster, and which has used advanced SAS®
Enterprise Content Categorization to establish 26,000 text rules for word
recognition and accurate classification. PSA incorporates effective data
infrastructure building, report linking, fast information searches, and
diagnosis and forecasting of enterprise problems.
10:00 a.m.
Creating Interval Target Scorecards with Credit Scoring
for SAS® Enterprise Miner™
Miguel Maldonado, SAS
Wendy Czika, SAS
Susan Haller, SAS
Naeem Siddiqi, SAS
Paper 094-2013
Credit Scoring for SAS® Enterprise Miner™ has been widely used to develop
binary target probability of default scorecards, which include scorecards for
application and behavior scoring. Extending the weight-of-evidence
binned scorecard methodology to interval-based targets such as loss given
default and exposure at default is a natural step in the evolution of
scorecard development methodologies. This paper discusses several new
methods available in Credit Scoring for SAS Enterprise Miner that help build
scorecards based on interval targets. These include cutoffs that convert the
interval target to a binary form and an algorithm that maintains the
continuous nature of the interval target. Each method is covered in detail,
and all methods are compared to determine how they perform against
known target values.
10:30 a.m.
Variable Reduction in SAS® by Using Weight of Evidence
and Information Value
Alec Lin, PayPal, a division of eBay
Paper 095-2013
Variable reduction is a necessary and crucial step in accelerating model
building without losing potential predictive power. This paper provides a
SAS® macro that computes weight of evidence and information value for all
potential predictors at the beginning stage of modeling. The SAS output
generated at the end of the program will rank continuous, ordinal, and
categorical variables by their predictive power, which can lend useful
insights to variable reduction. The reduced list of variables enables
statisticians to quickly identify the most informative variables for building
logistic regression models.
11:00 a.m.
Incremental Response Modeling Using SAS® Enterprise
Miner™
Taiyeong Lee, SAS
Ruiwen Zhang, SAS
Laura Ryan, SAS
Xiangxiang Meng, SAS
Paper 096-2013
Direct marketing campaigns that use conventional predictive models target
all customers who are likely to buy. This approach can lead to wasting
money on customers who will buy regardless of the marketing contact.
However, incremental response models that use a pair of training data sets
(treatment and control) measure the incremental effectiveness of direct
44
www.sasglobalforum.org/2013
marketing. These models look for customers who are likely to buy or
respond positively to marketing campaigns when they are targeted but are
not likely to buy if they are not targeted. The revenue generated from those
customers is called incremental revenue. This paper shows how to find that
profitable customer group and how to maximize return on investment by
using SAS® Enterprise Miner™.
11:30 a.m.
Estimates of Personal Revenue from Credit and
Sociodemographic Information Combining Decision
Trees and Artificial Neural Networks (ANN)
Deybis Florez Hormiga, Colpatria Bank
Paper 097-2013
In different bank processes, a typical problem is determining customer
revenue. Customer revenue information is very important and highly
impacts these processes. As a result, finding a method to estimate the
revenue of customers for validation, segmentation, profiling, business
strategies, risk mitigation, regulatory compliance, or simply as information
is critical. Due to the amount of information and the high volatility of the
income reported by different clients, SEMMA methodology is used with
SAS® Enterprise Miner™. Starting from a fine segmentation using decision
trees and then Artificial Neural Networks (ANN) in each of the segments,
higher performance is achieved by including credit information and
customer sociodemographic variables.
1:30 p.m.
Where Should I Dig? What to Do before Mining Your
Data
Stephanie Thompson, Datamum
Paper 098-2013
Data mining involves large amounts of data from many sources. In order to
successfully extract knowledge from data, you need to do a bit of work
before running models. This paper covers selecting your target and data
preparation. You want to make sure you find gold nuggets and not pyrite.
The work done up front will make sure your panning yields results and is
not just a trip down an empty shaft.
2:00 p.m.
Relate, Retain, and Remodel: Creating and Using
Context-Sensitive Linguistic Features in Text Mining
Models
Russell Albright, SAS
Janardhana Punuru, SAS
Lane Surratt, SAS
Paper 100-2013
Text mining models routinely represent each document with a vector of
weighted term frequencies. This bag-of-words approach has many
strengths, one of which is representing the document in a compact form
that can be used by standard data mining tools. However, this approach
loses most of the contextual information that is conveyed in the
relationship of terms from the original document. This paper first teaches
you how to write pattern-matching rules in SAS® Enterprise Content
Categorization and then shows you how to apply these patterns as a
parsing step in SAS® Text Miner. It also provides examples that improve on
both supervised and unsupervised text mining models.
Data Mining and Text Analytics — Room 3016
2:30 p.m.
Replacing Manual Coding of Customer Survey
Comments with Text Mining: A Story of Discovery with
Textual Data in the Public Sector
Jared Prins, Alberta Tourism, Parks and Recreation
(Invited) Paper 099-2013
A common approach to analyzing open-ended customer survey data is to
manually assign codes to text observations. Basic descriptive statistics of
the codes are then calculated. Subsequent reporting is an attempt to
explain customer opinions numerically. While this approach provides
numbers and percentages, it offers little insight. In fact, this method is
tedious and time-consuming and can even misinform decision makers. As
part of the Alberta government’s continual efforts to improve its
responsiveness to the public, the Alberta Parks division transitioned from
manual categorization of customer comments to a more automated
method that uses SAS® Text Miner. This switch allows for faster analysis of
unstructured data, and results become more reliable through the
consistent application of text mining.
Data Mining and Text Analytics — Room 2004
3:30 p.m.
Data Mining of U.S. Patents: Research Trends of Major
Technology Companies
Ken Potter, SAIC
Robert Hatton, SAIC
Paper 101-2013
Research initiatives are normally closely held corporate secrets. Insights into
research trends are difficult to extract from public information, but data
mining of the U.S. Patent and Trademark Office (USPTO) patent grants
provides an opportunity to expose interesting trends and areas of interest
as indicated by activity in related patent areas. This paper covers assessing
the vast USPTO information repository and the analytical methodology that
extracts patent grant information from multiple formats and produces
interesting insights into research trends for several major technology
companies.
4:00 p.m.
A Tale of Two SAS® Technologies: Generating Maps of
Topical Coverage and Linkages in SAS User Conference
Papers
Denise Bedford, Kent State University
Richard La Valley, Strategic Technology Solutions
Barry deVille, SAS
Paper 102-2013
This paper discusses how SAS® technologies -- Text Analytics and Content
Categorization Suite -- were used to generate comprehensive and dynamic
summaries of the entire corpus of SAS user presentations from inception to
the present. The goal was to improve access to the conference proceedings
for SAS users and conference attendees in particular. The research
addresses two important access points to conference papers -- Industry
Solutions and Technology Solutions. The findings of this research suggest
that both suites are powerful tools that can be used in complementary or
independent approaches to generate similar results. The Industry Solution
perspective generated by both technologies surfaced common access
points. The Technology Solution perspectives also generated similar
perspectives when comparable rule sets were leveraged.
4:30 p.m.
Unleashing the Power of Unified Text Analytics to
Categorize Call Center Data
Saratendu Sethi, SAS
Jared Peterson, SAS
Arila Barnes, SAS
Paper 103-2013
Business analysts often want to take advantage of text analytics to analyze
unstructured data. With that in mind, SAS is delivering a new web-based
application that is designed to put the power of SAS® Text Analytics into
the hands of the analyst. This application brings together the power of SAS®
Text Miner and SAS® Content Categorization in a single user interface that
enables users to automatically create statistical and rule-based models
based on their domain knowledge. This paper demonstrates how a
business analyst in a call center environment can identify emerging topics,
generate automatic rules for those topics, edit and refine those rules to
improve results, derive insights through visualization, and deploy the
resulting model to score new data.
5:00 p.m.
Deciphering Emoticons for Text Analytics: A MacroBased Approach
Chad Atkinson, Sinclair Community College
Paper 104-2013
Emoticons, initialisms, and acronyms can evade routine processing, and the
difference between "this was a great class :)" and "ZOMG that was the best
class ever :-7" might be significant. This paper develops a macro that
converts select emoticons, initialisms, and acronyms to text that can be
parsed by SAS® Text Analytics or SAS® Sentiment Analysis Studio.
5:30 p.m.
Be Customer Wise or Otherwise: Combining Data Mining
and Interactive Visual Analytics to Analyze Large and
Complex Customer Resource Management (CRM) Data
Aditya Misra, Nanyang Technological University
Kam Tin Seong, Singapore Management University
Junyao Ji, SAS Institute
Paper 105-2013
In this competitive world, more and more companies, such as our project
sponsor, a global logistics company, are exploring the potential use of data
mining techniques to make informed and intelligent marketing strategies.
We conducted a market segmentation study using a comprehensive set of
customer transaction and profile data. This paper aims to report on our
experience gained in using the interactive visual analytics and data mining
techniques of JMP® to perform customer segmentation analysis. We share
our views on how interactive visual analytics and data mining techniques
can empower everyday data analysts to gain useful insights and formulate
informed decisions by demonstrating the interactive data visualization
techniques of JMP such as graph builder, parallel plots, and bubble plots.
www.sasglobalforum.org/2013
45
Financial Services — Room 2010
Foundations and Fundamentals — Room 2008
1:30 p.m.
8:00 a.m.
How to Improve Your Risk-Adjusted Return on Capital:
Pricing Optimization in Lending
Macro Basics for New SAS® Users
Boaz Galinson, LEUMI
(Invited) Paper 106-2013
Lending is the core business of most banks. One may think that a bank
should opt for a lending pricing strategy of seeking the highest price that
the credit officer can obtain from his borrower. This paper claims that
following a "maximal price" strategy will eventually lead to an inferior credit
portfolio. I describe how to price a loan to meet at least the return required
by the stock holders and to improve RAROC. The strategy may be a
challenge in some assets classes. It can be difficult to agree on a price which
includes the minimal credit risk premium which compensates the risks. A
solution which accounts for all borrower activities with the bank is
presented.
2:30 p.m.
n Ounce of Prevention Is Worth a Pound of Cure: How
SAS® Helps Prevent Financial Crime with an Analytical
Approach to Customer Due Diligence
Scott Wilkins, SAS
Paper 107-2013
The increasing regulatory expectations on the risk rating of high-risk clients
and the emphasis on identification of foreign relationships with existing
customers has driven financial institutions to enhance their Customer Due
Diligence (CDD) processes. This paper outlines how organizations leverage
SAS® to deploy on-boarding and ongoing Customer Due Diligence
programs. It will explore analytical techniques for risk ranking customers,
best practices for deploying these programs, as well as how the SAS
approach incorporates a proactive, analytically driven triggering of new
investigations based on detected customer events or behavior. SAS can
provide organizations "The Power to Know" their customers and the risk
they may represent to their financial institutions.
3:30 p.m.
Next-Generation Detection Engine for Fraud and
Compliance
Ryan Schmiedl, SAS
Paper 108-2013
SAS’ next-generation approach provides a pivotal shift in how financial
institutions assess and govern customer risk. This paper discusses how
companies can aggregate, sum and understand patterns on huge volumes
of data; run more proactive what-if scenarios to identify and focus efforts
on the most critical investigations; and understand the impacts and
opportunity costs across scenarios.
46
www.sasglobalforum.org/2013
Cynthia Zender, SAS
Paper 120-2013
Are you new to SAS®? Do you look at programs written by others and
wonder what those & and % signs mean? Are you reluctant to change code
that has macro variables in the program? Do you need to perform repetitive
programming tasks and don’t know when to use DO versus %DO? This
paper provides an overview of how the SAS macro facility works and how
you can make it work in your programs. Concrete examples answer these
and other questions: Where do the macro variables live? What does it mean
when I see multiple ampersands (&&)? What is a macro program, and how
does it differ from other SAS programs? What’s the big difference between
DO and IF and %DO and %IF?
9:00 a.m.
Reading Data from Microsoft Word Documents: It's
Easier Than You Might Think
John Bentley, Wells Fargo Bank
Paper 121-2013
SAS® provides the capability of reading data from a Microsoft Word
document, and it's easy once you know how to do it. Using the FILENAME
statement with the DDE engine makes it unnecessary to export to Excel or
work with an XML map. This paper goes through the steps of having SAS
read a Word document and shares a live example that demonstrates how
easy it is. All levels of SAS users may find this paper useful.
9:30 a.m.
Do Not Let a Bad Date Ruin Your Day
Lucheng Shao, University of California at Irvine
Paper 122-2013
Just as we have to step out of fairy-tale land and into reality when we grow
up, we can’t always expect the input dates to be good. This paper shows
you what SAS® does when it runs into input dates that are normally good
but have now gone bad, and how those problems can be addressed by
code. It is intended for readers who are familiar with Base SAS but not with
bad dates.
10:00 a.m.
PROC DATASETS: The Swiss Army Knife of SAS®
Procedures
Michael Raithel, Westat
(Invited) Paper 123-2013
This paper highlights many of the major capabilities of PROC DATASETS. It
discusses how it can be used as a tool to update variable information in a
SAS data set; provide information on data set and catalog contents; delete
data sets, catalogs, and indexes; repair damaged SAS data sets; rename
files; create and manage audit trails; add, delete, and modify passwords;
add and delete integrity constraints; and more. The paper contains
examples of the various uses of PROC DATASETS that programmers can cut
and paste into their own programs as a starting point. After reading this
paper, a SAS programmer will have practical knowledge of the many
different facets of this important SAS procedure.
11:00 a.m.
The SAS® Programmer's Guide to XML and Web Services
Chris Schacherer, Clinical Data Management Systems, LLC
Paper 124-2013
Because of XML's growing role in data interchange, it is increasingly
important for SAS® programmers to become familiar with SAS technologies
and techniques for creating XML output, importing data from XML files, and
interacting with web services -- which commonly use XML file structures for
transmission of data requests and responses. The current work provides
detailed examples of techniques you can use to integrate these data into
your SAS solutions using SAS® XML Mapper, the XML LIBNAME engine, the
Output Delivery System, the FILENAME statement, and new SOAP functions
available beginning in SAS 9.3.
1:30 p.m.
Essentials of the Program Data Vector (PDV): Directing
the Aim to Understanding the DATA Step
Arthur Li, City of Hope
(Invited) Paper 125-2013
Beginning programmers often focus on learning syntax without
understanding how SAS® processes data during the compilation and
execution phases. SAS creates a new data set, one observation at a time,
from the program data vector (PDV). Understanding how and why each
automatic or user-defined variable is initialized and retained in the PDV is
essential for writing an accurate program. Among these variables, some
variables deserve special attention, including variables that are created in
the DATA step, by using the RETAIN or the SUM statement, and via BYgroup processing (FIRST.VARIABLE and LAST.VARIABLE). In this paper, you
are exposed to what happens in the PDV and how these variables are
retained from various applications.
2:30 p.m.
The Magnificent DO
Paul Dorfman, Dorfman Consulting
debuggers are good programmers. This paper covers common problems
including missing semicolons and character-to-numeric conversions, and
the tricky problem of a DATA step that runs without suspicious messages
but, nonetheless, produces the wrong results. For each problem, the
message is deciphered, possible causes are listed, and how to fix the
problem is explained.
4:30 p.m.
30 in 20 Things You May Not Know about SAS®
Tim Berryhill, Wells Fargo
Paper 128-2013
30 things you may not know SAS® can do. In 20 minutes, I hope to widen
your eyes and improve your programming. I have used SAS on many
platforms and operating systems, with many databases. Most of these ideas
will run anywhere SAS runs.
5:00 p.m.
Hashing in PROC FCMP to Enhance Your Productivity
Donald Erdman, SAS
Andrew Henrick, SAS
Stacey Christian, SAS
Paper 129-2013
Hashing has been around in the DATA step since 2002 starting with SAS® 9.
Hashing is used mainly to improve performance for activities like merging
and searching. Starting in SAS® 9.3, hashing functionality is now available in
user-defined subroutines through PROC FCMP. While subroutines already
encapsulate and modularize the code to make programs reusable, with the
addition of hashing, now users can extend the scope of their program and
tackle larger problems without sacrificing simplicity. The complete hashing
syntax supported in PROC FCMP will be outlined, as well as how it differs
from the DATA step. Examples will also be provided demonstrating just
how hashing in user-defined subroutines can be utilized to improve
performance and streamline an existing program.
(Invited) Paper 126-2013
Hands-on Workshops — Room 2011
Any high-level computer program can be written using just three
fundamental constructs: Sequence, Selection, and Repetition. The latter
forms the foundation of program automation, making it possible to execute
a group of instructions repeatedly, modifying them from iteration to
iteration. In SAS® language, explicit repetition is implemented as a standalone structural unit via the DO loop - a powerful construct laden with a
myriad of features. Many of them still remain overshadowed by the
tendency to structure code around the implied loop - even when it makes
the program more complex or error-prone. We will endeavor to both
straighten out some such incongruities and give the sense of depth and
breadth of the magnificent SAS construct known as the DO loop.
8:00 a.m.
3:30 p.m.
SAS® Workshop: SAS® Add-In for Microsoft Office 5.1
Eric Rossland, SAS
Paper 526-2013
This workshop provides hands-on experience using the SAS® Add-In for
Microsoft Office. Workshop participants will:
• access and analyze data
• create reports
• use the SAS add-in Quick Start Tools
Errors, Warnings, and Notes (Oh My): A Practical Guide
to Debugging SAS® Programs
Susan Slaughter, Avocet Solutions
Lora Delwiche, Univeristy of California, Davis
(Invited) Paper 127-2013
This paper is based on the belief that debugging your programs is not only
necessary, but also a good way to gain insight into how SAS® works. Once
you understand why you got an error, a warning, or a note, you'll be better
able to avoid problems in the future. In other words, people who are good
www.sasglobalforum.org/2013
47
Hands-on Workshops — Room 2020
8:00 a.m.
• register a user in the metadata
• manage access to application features with roles
Introduction to ODS Graphics
Hands-on Workshops — Room 2020
(Invited) Paper 141-2013
10:00 a.m.
Chuck Kincaid, Experis Business Intelligence and Analytics
SAS® has a new set of graphics procedures called ODS Graphics. They are
built on the Graph Template Language (GTL) in order to make the powerful
GTL easily available to the user. PROC SGPLOT and PROC SGPANEL are two
of the procedures that can be used to produce powerful graphics that
previously required a lot of work. This upgrade is similar to the ease-of-use
upgrade in output manipulation when ODS was first published. This handson workshop teaches you how to use PROC SGPLOT and PROC SGPANEL
and the new capabilities they provide beyond the standard plot. By using
these new capabilities, anyone can tell the story better.
Hands-on Workshops — Room 2024
8:00 a.m.
How to Use ARRAYs and DO Loops: Do I DO OVER or Do I
DO i?
Jennifer Waller, Georgia Health Sciences University
(Invited) Paper 140-2013
Do you tend to copy DATA step code over and over and change the
variable name? Do you want to learn how to take those hundreds of lines of
code that do the same operation and reduce them to something more
efficient? Then come learn about ARRAY statements and DO loops,
powerful and efficient data manipulation tools. This workshop covers when
ARRAY statements and DO loops can and should be used, how to set up an
ARRAY statement with and without specifying the number of array
elements, and what type of DO loop is most appropriate to use within the
constraints of the task you want to perform. Additionally, you will learn how
to restructure your data set using ARRAY statements and DO loops.
Hands-on Workshops — Room 2011
9:00 a.m.
Some Techniques for Integrating SAS® Output with
Microsoft Excel Using Base SAS®
Vince DelGobbo, SAS
Paper 143-2013
This paper explains some techniques to integrate your SAS® output with
Microsoft Excel. The techniques that are presented in this paper require
only Base SAS® 9.1 or above software, and can be used regardless of the
platform on which SAS software is installed. You can even use them on a
mainframe! Creating and delivering your workbooks on demand and in real
time using SAS server technology is discussed. Although the title is similar
to previous papers by this author, this paper contains new and revised
material not previously presented.
Hands-on Workshops — Room 2024
10:00 a.m.
A Row Is a Row Is a Row, or Is It? A Hands-On Guide to
Transposing Data
Christianna Williams, Independent Consultant
(Invited) Paper 142-2013
Sometimes life would be easier for the busy SAS® programmer if
information stored across multiple rows were all accessible in one
observation, using additional columns to hold that data. Sometimes it
makes more sense to turn a short, wide data set into a long, skinny one—
convert columns into rows. Base SAS® provides two primary methods for
converting rows into columns or vice versa: PROC TRANSPOSE and the
DATA step. How do these methods work? Which is best suited to different
transposition problems? The purpose of this hands-on workshop is to
demonstrate various types of transpositions using the DATA step and to
unpack the TRANSPOSE procedure. Afterward, you should be the office goto gal/guy for reshaping data.
SAS® Workshop: SAS® Enterprise Guide® 5.1
Eric Rossland, SAS
Paper 527-2013
This workshop provides hands-on experience using SAS® Enterprise Guide®.
Workshop participants will:
• access different types of data
• analyze data using the Data Explorer
• create reports and analyses
10:00 a.m.
SAS® Workshop: SAS® Platform Administration
Christine Vitron, SAS
Paper 528-2013
This workshop provides hands-on experience using SAS® Management
Console to administer the platform for SAS® Business Analytics. Workshop
participants will:
48
www.sasglobalforum.org/2013
Hands-on Workshops — Room 2011
11:00 a.m.
SAS® Workshop: SAS® Visual Analytics 6.1
Eric Rossland, SAS
Paper 529-2013
This workshop provides hands-on experience with SAS® Visual Analytics.
Workshop participants will:
• explore data with SAS® Visual Analytics Explorer
• design reports with SAS® Visual Analytics Designer
1:30 p.m.
• access different types of data
SAS® Workshop: Creating SAS® Stored Processes
• create reports and analyses
Eric Rossland, SAS
Paper 530-2013
This workshop provides hands-on experience creating SAS® Stored
Processes. Workshop participants will:
• use SAS® Enterprise Guide® to access and analyze data
• analyze data using the Data Explorer
3:30 p.m.
SAS® Visual Analytics 6.1
Eric Rossland, SAS
• create stored processes which can be shared across the organization
Paper 532-2013
• access the new stored process from the SAS® Add-In for Microsoft Office
This workshop provides hands-on experience with SAS® Visual Analytics.
Workshop participants will:
Hands-on Workshops — Room 2020
1:30 p.m.
Know Thy Data: Techniques for Data Exploration
Andrew Kuligowski, HSN
Charu Shankar, SAS Institute Toronto
(Invited) Paper 145-2013
Get to know the #1 rule for data specialists: Know thy data. Is it clean? What
are the keys? Is it indexed? What about missing data, outliers, and so on?
Failure to understand these aspects of your data will result in a flawed
report, forecast, or model. In this hands-on workshop, you learn multiple
ways of looking at data and its characteristics. You learn to leverage PROC
MEANS and PROC FREQ to explore your data, and how to use PROC
CONTENTS and PROC DATASETS to explore attributes and determine
whether indexing is a good idea. And you learn to employ powerful PROC
SQL’s dictionary tables to explore and even change aspects of your data.
Hands-on Workshops — Room 2024
1:30 p.m.
• explore data with SAS® Visual Analytics Explorer
• design reports with SAS® Visual Analytics Designer
Hands-on Workshops — Room 2020
3:30 p.m.
Hands-On SAS® Macro Programming Tips and
Techniques
Kirk Paul Lafler, Software Intelligence Corporation
(Invited) Paper 146-2013
The SAS® macro language is a powerful tool for extending the capabilities
of SAS. This hands-on workshop presents numerous tips and tricks related
to the construction of effective macros through the demonstration of a
collection of proven macro language coding techniques. Attendees learn
how to process statements containing macros; replace text strings with
macro variables; generate SAS code using macros; manipulate macro
variable values with macro functions; handle global and local variables;
construct arithmetic and logical expressions; interface the macro language
with the DATA step and SQL procedure; store and reuse macros;
troubleshoot and debug macros; and develop efficient and portable macro
language code.
Getting Started with the SAS/IML® Language
Rick Wicklin, SAS
Paper 144-2013
Do you need a statistic that is not computed by any SAS® procedure? Reach
for the SAS/IML® language! Many statistics are naturally expressed in terms
of matrices and vectors. For these, you need a matrix-vector language. This
hands-on workshop introduces the SAS/IML language to experienced SAS
programmers who are familiar with elementary linear algebra. The
workshop focuses on statements that create and manipulate matrices, read
and write data sets, and control the program flow. Learn how to write userdefined functions, interact with other SAS procedures and recognize
efficient programming techniques. Programs will be written using the SAS/
IML® Studio development environment. This course covers chapters 2-4 of
“Statistical Programming with SAS/IML Software” (Wicklin 2010).
Hands-on Workshops — Room 2011
2:30 p.m.
Hands-on Workshops — Room 2024
3:30 p.m.
Taking Full Advantage of sasCommunity.org: Your SAS®
Site
Don Henderson, Henderson Consulting Services
(Invited) Paper 147-2013
sasCommunity.org, a clearinghouse for technical information related to the
use of SAS® software, is managed and run by SAS users; free and open to all
SAS users to browse; contributed to by any SAS user once they create an ID;
and built on top of the same software as Wikipedia. So if you know how to
use Wikipedia, you have a head start on using sasCommunity.org. Learn
how to navigate the information contained on the site; discover the wealth
of its hidden treasures; and make even small contributions to enhance the
site for everyone. Find out the number of ways you can contribute, and
discover how you too can quickly make a difference in this worldwide
community of SAS users.
SAS® Workshop: SAS® Enterprise Guide® 5.1
Eric Rossland, SAS
Paper 531-2013
This workshop provides hands-on experience using SAS® Enterprise Guide®.
Workshop participants will:
www.sasglobalforum.org/2013
49
Hands-on Workshops — Room 2011
9:30 a.m.
4:30 p.m.
Estimating Patient Adherence to Medication with
Electronic Health Records Data and Pharmacy Claims
Combined
SAS® Workshop: SAS® Platform Administration
Christine Vitron, SAS
Paper 533-2013
This workshop provides hands-on experience using SAS® Management
Console to administer the platform for SAS® Business Analytics. Workshop
participants will:
• register a user in the metadata
• manage access to application features with roles
Pharma and Health Care — Room 2000
8:00 a.m.
Clinician Prescribing Feedback Site: Comparing Clinician
Prescribing Habits and Providing Actionable Patient
Lists
Michael Nash, Kaiser Permanente
Paper 165-2013
Which doctor is prescribing the most non-formulary medications? Which
patients are on a brand drug when an equivalent generic is available? These
questions and many more can be answered when using the Pharmacy
Feedback Site. This secure intranet site at Kaiser Northwest uses SAS/
GRAPH® HBAR and VBAR charts to compare clinician prescribing habits. Drill
down to compare all clinics, or all departments, or all doctors within a Clinic
or Specialty. Drill down even further to find patient lists so pharmacists or
clinician staff can perform outreach to members. The Pharmacy Feedback
Site also tracks costs and patients month to month. This paper shows you
how to create linked HTML files by using PROC GCHART and the HTML=
option.
8:30 a.m.
Identifying and Addressing Post-Marketing
Pharmaceutical Safety Surveillance and Spontaneous
Reported Events
Carrie Boorse, SAS
Kathy Schaan, SAS
Stuart Levine, SAS
Paper 166-2013
While pharmaceutical medications and medical devices must undergo
clinical trials to determine their safety, often their long-term side effects will
not be recognized until the medication or device has been approved and
consumed by a larger population of patients for a longer period of time.
Signals that an adverse event is brewing need to be identified by using
post-marketing data as expeditiously as possible. Using SAS® analytical
platform and foundation products, a solution was implemented that
incorporates recognized statistical procedures and enables users to
incorporate new and enhanced procedures for signal determination. The
solution also alerts users that a signal has been triggered, while also
managing the movements and results of the signal’s investigation.
50
www.sasglobalforum.org/2013
Beinan Zhao, Palo Alto Medical Foundation Research
Insitute
Eric Wong, Palo Alto Medical Foundation Research Institute
Latha Palaniappan, Palo Alto Medical Foundation
Paper 167-2013
Estimating patient adherence to medication is critical for comparative
effectiveness, patient-centered outcomes research, and epidemiological
studies. Using a comprehensive electronic medical record system (EpicCare)
that has been in practice for 11 years with more than one million patients,
the prevailing adherence metrics (for example, medication possession ratio
and proportion days covered) were evaluated. However, these metrics
cannot be evaluated when a patient does not fill a medication order
(primary non-adherent) or fills it only once (early stop). With just a little
more effort, additional clinical information can be incorporated from
electronic health records to obtain refined estimates of adherence. This
paper proposes a few composite metrics that might be of specific interest
to researchers and clinicians.
10:00 a.m.
Measuring Medication Adherence with Simple Drug Use
and Medication Switching
Stacy Wang, Walgreens
Zhongwen Huang, Walgreens
Seth Traubenberg, Walgreen Co.
Paper 168-2013
In this paper, we demonstrate SAS®-based solutions that allow providers to
calculate adherence across a range of prescribing patterns. The code
provided allows PDC to be calculated at both the therapeutic class level
and the patient disease level. Refining existing methodologies has
increased the efficiency of the calculations.
10:30 a.m.
SAS® Tools for Transparent and Reproducible Research:
Medication History Estimator
Brian Sauer, SLC VA Medical Center
Tao He, University of Utah
Paper 169-2013
The Medication History Estimator (MHE) is designed to output data at the
course-level; i.e., one row per drug course. A course and period proportion
of days covered (PDC) is calculated for each medication. Reports that
describe the frequency and percent of users for each medication product,
average duration of medication courses, medication possession ratios and
Kaplan-Meier based persistency curves are automatically generated and
formatted for professional reports and journal publications.
11:00 a.m.
3:00 p.m.
What Do Your Consumer Habits Say About Your Health?
Using Third-Party Data to Predict Individual Health Risk
and Costs
Evaluating System-Wide Process Improvement in a
Health-Care System: Data Through Analysis
Albert Hopping, SAS
Satish Garla, SAS
Rick Monaco, SAS Institute
Sarah Rittman, SAS
Paper 170-2013
The Affordable Care Act is bringing dramatic changes to the health care
industry. Previously uninsured individuals are buying health insurance and
consuming health care services differently. These changes are forcing
insurers to reevaluate marketing, engagement and product design
strategies. The first step in addressing these challenges is understanding
the financial risk of new entrants into the marketplace. How do you predict
the risk of a person without any historical cost information? What if all you
know is the name and address? The finance industry has long been using
third-party consumer data to predict future finance habits and credit risk.
This paper takes a look at applying advanced analytics from SAS to thirdparty data for predicting health care utilization risk and costs.
Pharma and Health Care — Room 3016
1:30 p.m.
To Infinity and Beyond: Current and Future State of Big
Data and Analytics in Life Sciences
Matthew Becker, inVentiv Health Clinical
(Invited) Paper 503-2013
Our biggest asset is our data. We have all heard a semblance of these words
in the Life Sciences industry. The questions many of us ask are: Are we
tapping into the data as we should? Are we pulling the multiple avenues of
data together with all the parameters that could be analyzed? Are we
providing analytics in an educational way to our end-user(s)? In this
keynote, we will look at the current state of big data in the Life Sciences
industry and share a glimpse into the future of big data and analytics.
Pharma and Health Care — Room 2000
2:30 p.m.
Moving to SAS® Drug Development 4.2
Magnus Mengelbier, Limelogic
Eric Wong, Palo Alto Medical Foundation Research Institute
Lubna Qureshi, Palo Alto Medical Foundation Research
Institute
Pragati Kenkare, Palo Alto Medical Foundation Research
Institute
Dorothy Hung, Palo Alto Medical Foundation Research
Institute
Paper 172-2013
Disruptive system changes are required for sustaining high-quality and
affordable health-care delivery systems. Successful, transformative healthcare system changes are few and even fewer have been rigorously
evaluated. Electronic health records and changes in health IT provide an
opportunity to leverage an explosion of data in measuring the impact of
process improvement initiatives. This paper provides an example of
assessing the impact of a system-wide change in a large, multi-specialty
health-care system serving two million patients with a 13-year history of
using electronic health records. Lessons from ETL all the way to statistical
analysis are detailed including relevant SAS® procedures.
Pharma and Health Care — Room 3016
3:30 p.m.
Modern SAS® Programming: Using SAS® Grid Manager
and SAS® Enterprise Guide® in a Global Pharmaceutical
Environment
David Edwards, Amgen
Greg Nelson, ThotWave Technologies, LLC.
(Invited) Paper 173-2013
Amgen, like most large biotechnology companies, uses SAS® to support the
drug discovery process. Equipped with a vision to fully leverage its global
workforce and to maximize its IT investments, Amgen developed a research
informatics infrastructure based on SAS to deliver value around the globe.
This paper will highlight many aspects of this project including business
justification, requirements, design, verification and validation, and
production migration for over 1500 programmers and statisticians spread
across three continents. We will highlight some of the challenges we faced
and how these were overcome using improved processes, modern
technologies such as SAS® Grid Manager and SAS® Enterprise Guide® and
the combined efforts of a global project team.
Paper 171-2013
Life Science organizations have a long investment into business processes,
standards, and conventions that make it difficult to simply turn to a new
generation of analysis environments. SAS® Drug Development 4.1
integrates many key features found in current analysis environments that
are spread across several applications and systems, which need to be
monitored and managed accordingly. The paper considers a set of SAS®
programs and how the SAS Drug Development repository, workspace, and
workflow features support a common business process with all of the
associated tools and utilities. The result is a short list of points to consider
and some tricks for moving a business process from a PC SAS or SAS server
environment to the new release of SAS Drug Development.
Pharma and Health Care — Room 2000
4:30 p.m.
Medical Versus Pharmacy Insurance: Which Is More CostEffective for Providing the Prescription? Solving the
Problem Via SAS® Enterprise Guide®.
Amber Schmitz, Prime Therapeutics
Paper 174-2013
Data-driven decisions that provide strategic solutions. These are buzzwords
and phrases we have all heard before, but actually applying those words to
deliver actionable data is less commonplace than it should be. This paper
www.sasglobalforum.org/2013
51
explores how to use both pharmacy and medical insurance claims data in
order to assess drug utilization behavior across the medical and pharmacy
insurance benefits. The final result provides actionable data to assess
moving drug fills to the more cost-effective insurance benefit. This paper
explores: 1) Writing programs for efficient data pulls, 2) Macrotizing
program code to allow for flexible analysis constraints, 3) Using SAS®
Enterprise Guide® Tasks for analysis, and 4) Demonstrating business
intelligence via built-in graph options.
5:00 p.m.
Employee Wellness Programming Using SAS® Enterprise
Guide®
Yehia Khalil, Norton Healthcare
Tina Hembree, Norton Healthcare
Sandra Brooks, Norton Healthcare
Paper 175-2013
More businesses are using employee wellness programs to improve the
health of their employees (improve productivity levels, reduce
absenteeism, and reduce disability claims) while at the same time reducing
health care costs. The success of any wellness program depends on two
main rudiments. One: identify factors that drive up health care costs in the
organization such as smoking, obesity, chronic conditions, and others. Two:
achieve adequate employee engagement level in wellness programs and
identify barriers to achieving this level. The real challenge for any wellness
program is to incorporate the different data sources such as health risk
assessments (HRAs), demographics, medical claims data, and focus group
reports to build a comprehensive wellness program that considers the
various needs of the business population.
Planning and Support — Room 2010
9:00 a.m.
Managing and Monitoring Statistical Models
Nate Derby, Stakana Analytics
(Invited) Paper 190-2013
Managing and monitoring statistical models can present formidable
challenges when you have multiple models used by a team of analysts over
time. How can you efficiently ensure that you're always getting the best
results from your models? In this paper, we'll first examine these challenges
and how they can affect your results. We'll then look into solutions to those
challenges, including lifecycle management and performance monitoring.
Finally, we'll look into implementing these solutions both with an in-house
approach and with SAS® Model Manager.
10:00 a.m.
SAS® Certification: Understand the Benefits of SAS
Certification, Which SAS Certifications Are Available,
and What SAS Certification Can Do for You
Andrew Howell, ANJ Group Pty Ltd
Paper 191-2013
SAS® has long had certification available for its programming language and
for its flagship data mining product, SAS® Enterprise Miner™. More recently
with the release of the SAS®9 platform suite have come certifications in
SAS® Data Integration Studio, SAS® Business Intelligence, and SAS® Platform
Administration. But what are the benefits (and some of the misconceptions)
of SAS Certification? What is available, and what's in it for organizations,
their staff and for SAS consultants to become SAS Certified?
10:30 a.m.
8:00 a.m.
The Successful SAS® Shop: 10 Ideas, Suggestions, and
Radical Notions
Communicating Standards: A Code Review Experience
Paper 192-2013
David Scocca, Rho, Inc.
Paper 188-2013
We need ways to pass along good programming practices. All but the
smallest companies will have programmers with varying levels of tenure
and experience. Standards and best practices change, but in a deadlinedriven world, we re-use old programs with minimal revision. Programmers
develop habits and can be slow to incorporate new approaches that might
simplify code or improve performance. We developed and rolled out an inhouse code review process to address these issues. This paper reports our
strategy for promoting and performing the reviews and describes the
results.
8:30 a.m.
If You Have Programming Standards, Please Raise Your
Hand: An Everyman's Guide
Dianne Louise Rhodes, US Census Bureau
Paper 189-2013
This paper goes through a step-by-step process of developing
programming standards, classifying them, and entering them into a
database. This database can then be used to develop style sheets and check
lists for peer review and testing. Through peer reviews and in preparation
for them, programmers learn good programming practices. We describe in
detail the most common standards, and why and how they should be
applied.
52
www.sasglobalforum.org/2013
Lisa Horwitz, SAS
A SAS® shop might have as few as one or two programmers or as many as
several thousand programmers, analysts, statisticians, data stewards,
administrators, and help desk monitors. These SAS shops might be very
new to their organizations, or they might have grown and evolved over a
period of many years. Regardless of their size, age, or overall mission, there
are some common factors that allow these groups of people to find
satisfaction and reward in what they do. This paper details 10 ideas,
suggestions, and radical notions to ensure happy and productive SAS
programmers.
11:30 a.m.
Creating an Interactive SAS® E-Textbook with iBooks
Author for the iPad
William Zupko, U. S. Census Bureau
Paper 193-2013
Mobile media is becoming more popular and prevalent in today's
workplace. Even though few apps on the iPad actively run SAS® programs,
the iPad can be utilized as a teaching tool and reference database. iBooks
Author allows for the creation of interactive textbooks from anybody that
allow users to learn SAS in a self-paced environment. Widgets allow
screenshots to show how programs are run and use reviews to check
comprehension. These widgets also allow the inclusion of Keynote slides.
These interactive textbooks are especially excellent for SAS conferences, as
the text can be applied directly and include PowerPoint presentations,
creating a mobile library that can be used in an easily accessible format,
perfect for reference or training.
Quick Tips — Room 2003
8:00 a.m.
Automating the Flow of Presentations in Coder's Corner
or Quick Tips
Erik Tilanus, Synchrona
Paper 292-2013
The Quick Tips section (former Coder's Corner) is characterized by a rapid
flow of many short presentations. Reading bios and starting presentations
by hand is slowing down this flow. So we use SAS® to automate this flow.
8:15 a.m.
You’ve Got SASMAIL! A Simple SAS® Macro for Sending
e-Mails
Rajbir Chadha, Cognizant Technology Solutions
Paper 340-2013
This paper talks about a way to have SAS® send out automatic e-mails once
the process finishes and include the log output or reports as an attachment.
The SASMAIL custom macro function combines the functionality of
FILENAME EMAIL, SAS macros, and the SQL procedure to deliver the
intended results in a simplified way. The SASMAIL macro uses a lookup for
the user’s login and e-mail address. This allows the macro to work with
minimum or no arguments. The function allows users to customize what
they want to see in the e-mail, including the e-mail list, the attachment, and
even the e-mail content. Users can even include a Microsoft Excel file or a
summary report as the attachment.
8:45 a.m.
Adding Graph Visualization on SAS® ODS Output
Yu Fu, Oklahoma State University
Shirmeen Virji, Oklahoma State University
Goutam Chakraborty, Oklahoma State University
Miriam McGaugh, Oklahoma State University
Paper 310-2013
SAS® tools are normally used to produce statistical graphs such as pie
charts, bar charts, various plots, dashboards, and even geographical maps.
However, many SAS users may want to enhance their output by
incorporating various diagrams such as networking, cluster, and process
flows. In this paper, we will introduce a method to add specific graphs onto
SAS ODS output by interacting with Graphviz (an open source graph
visualization software) in Base SAS®.
9:00 a.m.
"How May I Help?" The SAS® Enterprise Guide® Analyze
Program Feature
Ramya Purushothaman, Cognizant Technology Solutions
Paper 311-2013
Have you ever been looking into a lengthy and complex SAS® code, maybe
inherited it or your own old program, and wished you could understand
what is happening inside without having to go through every line? Left
without any associated documentation and wondered where to start from?
Felt that it would be better to have a process flow representation of what
the code does, quickly? Then, the Analyze Program feature that SAS®
Enterprise Guide® offers might work for you! This paper discusses what to
expect of this feature and what not to with example analyses from the Life
Sciences industry.
9:15 a.m.
8:30 a.m.
Some Useful Utilities on UNIX Platform
Reporting Tips for No Observations
Paper 312-2013
Wuong Jodi Auyuen, Blue Cross Blue Shield Minnesota
Paper 320-2013
The goals for SAS developers to design applications include reporting
accurate information, delivering in a timely manner, meeting business
needs, and the presentation is easy to grasp. We design the report to meet
those goals and hopefully to cover potential questions. One of the
frequently asked questions is: I used to receive a session, say visitors from
Japan, why I don't see that session for the week of March 14, 2011? Even
though we don't need to code "No visitors from Japan due to Tsunami on
March 11, 2011", we could at least provide a generic message like "No data
returned for this session." so users won't be wondering whether they miss
the page or question the accuracy of the development work.
Kevin Chung, Fannie Mae
While using SAS® in UNIX platform, you might want to quickly browse data,
contents or the frequency of characteristic data fields in a SAS data set. You
can always write a SAS program and submit the program to get the results
you need. However, we are able to obtain this information in more efficient
and effective manner by using the UNIX shell scripts along with SAS codes.
This paper demonstrates some useful utilities in UNIX. This approach not
only saves your time but it also increases the productivity.
9:30 a.m.
PC and UNIX SAS® Reunited
Shiva Kalidindi, Amgen
Sarwanjeet Singh, Gerard Groups Inc.
Paper 313-2013
Have you ever wondered how you can use the best of PC and UNIX SAS
together and make a perfect world (well, almost perfect)? SAS/CONNECT®
allows you to use the Enhanced Editor and submit the code on UNIX. You
can submit one DATA step or PROC at a time, view the log in the Log
window as well as create data sets in the Work directory. It is a one-time
setup, and you do not have to compromise the PC Enhanced Editor ever.
This paper provides step-by-step instructions on how you can connect and
automate the PC-to-UNIX connection by using SAS. You will never have to
leave the Enhanced Editor again.
www.sasglobalforum.org/2013
53
9:45 a.m.
10:45 a.m.
SAS® Code to Make Excel Files Section 508 Compliant
Reordering Columns after PROC TRANSPOSE (or
Anytime You Want, Really)
Christopher Boniface, U.S. Census Bureau
Hung Pham, U.S. Census Bureau
Nora Szeto, U.S. Census Bureau
Paper 314-2013
Can you create hundreds of great looking Excel tables all within SAS® and
make them all Section 508 compliant at the same time? This paper will
examine how to use ODS tagsets, EXCELXP, and other Base SAS® features to
create fantastic-looking Excel worksheet tables that are all Section 508
compliant. This paper will demonstrate that there is no need for any
outside intervention or pre- or post-meddling with the Excel files to make
them Section 508 compliant. We do it all with simple Base SAS code.
10:00 a.m.
Reading an Excel Spreadsheet with Cells Containing Line
Endings
Larry Hoyle, IPSR, University of Kansas
Paper 315-2013
The creative ways people enter data into Excel spreadsheets can cause
problems when trying to import data into SAS® data sets. This paper
addresses the problem encountered when spreadsheet cells contain
multiple lines (that is, the cells have embedded line endings). Several
approaches to reading such data are described and compared.
10:15 a.m.
Maintaining Formats When Exporting Data from SAS®
into Microsoft Excel
Nate Derby, Stakana Analytics
Colleen McGahan, BC Cancer Agency
Paper 316-2013
Data formats often get lost when exporting from SAS® into Microsoft Excel
using common techniques such as PROC EXPORT or the ExcelXP tag set. In
this paper, we describe some tricks to retain those formats.
10:30 a.m.
Don’t Let the Number of Columns Hold You Back!
Douglas Liming, SAS
Paper 318-2013
Many databases have a column limit of approximately 2,000 columns.
Several SAS® PROCs, such as PROC NEURAL and PROC TRANSPOSE, produce
output that easily exceeds 2,000 columns. Here is a technique to code
around this business problem when using databases on the backend.
Maximize your columns using SAS to talk to the databases via multiple
tables, pull them together and split them back out.
Sau Yiu, Kaiser Permanente
Paper 319-2013
There are times when we want to rearrange the order of the columns in a
SAS® data set. This occurs most often after a PROC TRANSPOSE, when the
newly transposed columns do not appear in the order that we want. This
paper shows several methods which allow users to either sort the columns
by their names, or order the columns in any particular way.
11:00 a.m.
Wide-to-Tall: A Macro to Automatically Transpose Wide
SAS® Data into Tall SAS Data
James R Brown, Havi Global Solutions
Paper 321-2013
If your SAS® world involves forecasting or other date-specific data, you have
probably seen column names such as forecast_19224, sales_19230, or
inventory_19250. If several of these prefixes exist in a single file, the
underlying SAS data file could have thousands of columns. Analyzing this
data is an exercise in scrolling, note-taking, copying, and pasting. PROC
TRANSPOSE is not sophisticated enough to take on this challenge. This
paper presents a macro which will transform your data by automatically
creating a CSV file with distinct columns for the date, each prefix variable,
and any non-date-suffixed columns in your input. The non-wizardry behind
this makes use of the dictionary tables, SAS name lists (forecast_18950forecast_19049), and colon notation (forecast_:) to eliminate the task of
enumerating long lists of variable names.
11:15 a.m.
Nifty Tips For Data Change Tracking
Julie Kilburn, City of Hope
Rebecca Ottesen, City of Hope and Cal Poly State University,
San Luis Obispo
Paper 333-2013
Best practices for databases include keeping detailed audit trail information
about the data. These audit trail tables vary in complexity as well as size.
Generally speaking, the larger the database in tables (as well as in
observations), the larger the audit trail. We have discovered that leveraging
audit trail information in our automated reporting has been a huge
resource saver in terms of which observations need to be reprocessed for a
report. Even with minimal audit information (such as created by and
modified by dates at the data table level), automation processing time can
be greatly reduced by taking advantage of a new way of thinking and a few
handy SAS® functions.
11:30 a.m.
Programs? How to Process Your Inputs Faster
Jason Wachsmuth, Pearson
Paper 334-2013
This paper demonstrates how to pass multiple input files as macro variables
and run multiple SAS® programs in one batch. This technique uses
%INCLUDE statements, CALL SYMPUT, and SCAN functions in a control
program to avoid physically opening, running, and closing each program.
Implementing this style of programming replaces the bottleneck of
54
www.sasglobalforum.org/2013
defining %LET statements and enables you to process input files and
sequence dependent programs in proper order. Anyone who processes
data for a variety of routine operations will appreciate this solution.
efficiency improvement. This paper shows some sample code that divides
by zero and some benchmark results from changing the code to test for
zero denominators to avoid dividing by zero.
1:30 p.m.
2:30 p.m.
What's in a SAS® Variable? Get Answers with a V!
Running SAS® Programs Using Skype
Paper 322-2013
Paper 304-2013
If you need information on variable attributes, PROC CONTENTS will
provide all of the specifics. You could also access the SQL DICTIONARY
which contains tables filled with details on the variables in the active data
sets. If you need just one piece of information on a single variable, both of
these methods could prove to be cumbersome. However, SAS® has a whole
series of functions that can produce the information on one attribute of one
variable at one time. These functions are named with a V prefix followed by
descriptive term for the attribute. They include VNAME, VLABEL, VTYPE,
VFORMAT, and several others. We will demonstrate how these V functions
are useful not only in reporting but in standardizing the structure of a
database.
Skype is a well-known method used to talk to friends, family members, and
coworkers. It is one of the best applications available to make voice calls
over the Internet. In this paper we present a new, innovative way to use
SAS® with Skype. Here, we have developed a solution that allows users to
run SAS remotely through Skype. After installing the DLL from the API on
the application website, programmers can create scripts to control Skype.
By sending a specific message to a predefined user, programmers can
execute SAS on demand. This paper explains how to use Skype to run SAS
programs. It provides the Visual Basic script needed to communicate with
Skype and illustrates a real case scenario in which this technique is used.
William Murphy, Howard M Proskin & Assoc, Inc
1:45 p.m.
Quick and Easy Techniques for Fast Data Extraction
Mythili Rajamani, Kaiser Permanente
Deepa Sarkar, Kaiser Permanente
Jason Yang, Kaiser Permanente
Chris Greni, Kaiser Permanente
Paper 323-2013
Working with large data sets is a challenging and time-consuming job. This
paper tells some of the easy, useful data extraction tips and techniques to
reduce CPU usage to retrieve the specific data. The following subjects are
discussed in this paper: (1) creating a temporary table with the Key column
in the database (both DB2 and Teradata) and extracting the data from the
database (2) extracting data only for specific days, specific weeks (3)
automating the date parameter for repetitive and scheduled tasks.
Romain Miralles, Genomic Health
2:45 p.m.
An Overview of Syntax Check Mode and Why It Is
Important
Thomas Billings, Union Bank
Paper 327-2013
The syntax check options direct the SAS® system, when a syntax error
occurs while compiling source code, to enter a special mode to scan the
remainder of the job after the point where the error occurred, for syntax
errors. In this mode, only the header portion of some data sets are created,
permanent data sets are not replaced, but global commands are executed
(also a very few PROCs). The options controlling the mode are explained
and illustrated using simple test jobs. The effects of setting and resetting
the option within a job are explored, and there are some surprises along
the way. The risks of running with the options enabled vs. disabled are
discussed.
2:00 p.m.
3:00 p.m.
Dealing with Duplicates
Converting Thousands of Variables from Character to
Numeric: The One-Hour Fix
Christopher Bost, MDRC
Paper 324-2013
Variable values might be repeated across observations. If a variable is an
identifier, it is important to determine whether values are duplicated. This
paper reviews techniques for detecting duplicates with PROC SQL,
summarizing duplicates with PROC FREQ, and outputting duplicates with
PROC SORT.
2:15 p.m.
Not Dividing by Zero: The Last of the Low-Hanging
Efficiency Fruit
Bruce Gilsen, Federal Reserve Board
Wen Song, ICF International
Kamya Khanna, ICF International
Paper 328-2013
At the conclusion of many survey-based data collecting projects, recoding
the hundreds and thousands of character variables to “reserved scale”
specified numeric variables is a uncomplicated but cumbersome task for
SAS® programmers. If you are a person who likes to avoid a large amount of
typing as much as I do, this paper will give you an idea of how to maintain
high quality for this recoding task with minimal typing. This paper also
answers the following questions: How can you create a powerful SAS macro
that will write the IF-ELSE-THEN Statement for you? How can you avoid any
human errors such as typos? And how do you use Microsoft Excel to speed
up your work?
Paper 325-2013
As SAS® Institute has improved the efficiency of its code, some of the old
ways for users to improve efficiency, such as using WHERE or WHERE=
instead of IF in the DATA step, no longer make much difference. Current
user efforts to improve efficiency tend to focus on more sophisticated
techniques such as indexes or hashing. However, one of the classic
methods, not dividing by zero in the DATA step, can still offer a large
www.sasglobalforum.org/2013
55
3:15 p.m.
4:15 p.m.
Resources for Getting the 2010 US Census Summary
Files into SAS®
Checking Out Your Dates with SAS®
Rebecca Ottesen, City of Hope and Cal Poly State University,
San Luis Obispo
Paper 329-2013
At first glance, accessing the 2010 US Census data with SAS® seems like a
daunting task. The main limitation is that for the 2010 summary files it
seems that the Census has gravitated toward supporting data access via
Microsoft Access rather than SAS as they did in the past. However, there are
several tactics that can be deployed to make accessing this data with SAS
much easier. A thorough understanding of the Census Summary File data
structure and documentation can be used to leverage both SAS code from
programs that the Census previously supported and Census 2010 versioned
SAS programming available through other public sources. Knowledge of
the available resources can assist SAS analysts in taking advantage of this
rich data set.
3:30 p.m.
Array, Hurray, Array: Consolidate or Expand Your Input
Data Stream Using Arrays
William Benjamin, Owl Computer Consultancy LLC
Paper 330-2013
You have an input file with one record per month, but need a file with one
record per year. But you cannot use PROC TRANSPOSE because other fields
need to be retained or the input file is sparsely populated. The techniques
shown here enable you to either consolidate or expand your output data
using arrays. Sorted files of data records can be processed as a unit using
"BY Variable" groups and building an array of records to process. This
technique allows access to all of the data records for a "BY Variable" group
and gives the programmer access to the first, last, and all records in
between at the same time. This will allow the selection of any data value for
the final output record.
Christopher Bost, MDRC
Paper 335-2013
Checking the quality of date variables can be a challenge. PROC FREQ is
impractical with a large number of dates. PROC MEANS calculates summary
statistics but displays results as SAS® date values. PROC TABULATE,
however, can calculate summary statistics and format the results as dates.
This paper reviews these approaches plus the STACKODS option in SAS® 9.3
that might make PROC MEANS the preferred method for checking out your
dates.
4:30 p.m.
We All Have Bad Dates Once in a While...
Randall Deaton, BlueCross BlueShield of Tennessee
Patrick Kelly, BlueCross BlueShield of Tennnessee
Paper 336-2013
Dates in a corporate data arena can be a dangerous liaison. The strain of
translating corporate date types to SAS® date types can be tricky to
navigate, let alone bringing a third-party date type into the mix. Adding
third wheel to your SAS dates can create a comedy of errors. An
experienced SAS programmer with knowledge of the SAS macros and a few
clever programming tricks can more easily resolve your SAS Dates from a
sticky situation to an orderly affair.
4:45 p.m.
Increase Your Productivity by Doing Less
Robert Virgile, Robert Virgile Associates, Inc.
Arthur Tabachneck, myQNA, Inc.
Xia Keshan, Chinese Financial electrical company
Joe Whitehurst, High Impact Technologies
Paper 517-2013
3:45 p.m.
10-Minute JMP®
George Hurley, The Hershey Company
Paper 331-2013
Heard of JMP®, but haven't had time to try it? Don't want to devote 50
minutes to a talk about software that you might not want to use? This is the
talk to you. In 10 minutes, you will learn some of the amazing visualization
and modeling features in JMP and how to use them. This talk will JMP-start
your JMP usage. When it's complete, we suspect you will want to attend
some of the longer talks, too.
4:00 p.m.
Cool Views
Elizabeth Axelrod, Abt Associates Inc.
Paper 332-2013
Looking for a handy technique to have in your toolkit? Consider SAS® Views,
especially if you work with large data sets. After a brief introduction to
Views, I will show you several cool ways to use them that will streamline
your code and save workspace.
56
www.sasglobalforum.org/2013
Using a keep dataset’ option when declaring a data option has mixed
results with various SAS procedures. It might have no observable effect
when running PROC MEANS or PROC FREQ but, if your datasets have many
variables, it could drastically reduce the time required to run some procs
like PROC SORT and PROC TRANSPOSE. This paper describes a fairly simple
macro that could easily be modified to use with any proc that defines which
variables should be kept and, as a result, make your programs run 12 to 15
times faster.
Reporting and Information Visualization — Room
2002
8:00 a.m.
Make a Good Graph
Sanjay Matange, SAS
Paper 361-2013
A graph is considered effective if the information contained in it can be
decoded quickly, accurately and without distractions. Rules for effective
graphics – developed by industry thought leaders such as Tufte, Cleveland
and Robbins – include maximizing data ink, removing chart junk, reducing
noise and clutter, and simplifying the graph. This presentation covers these
principles and goes beyond the basics, discussing other features that make
a good graph: the use of proximity for magnitude comparisons, nonlinear
9:00 a.m.
the names of the style attributes that you want to change. This presentation
provides concrete examples to illustrate how to use STYLE= overrides with
PRINT, REPORT, and TABULATE. As the examples move from simple to
complex, you learn how to change fonts, add text decoration, alter the
interior table lines, perform traffic-lighting, and insert images into your ODS
output files using some ODS magic to improve your reports.
A Beginner's Introduction to an Idiot's Guide to PROC
TEMPLATE and GTL for Dummies
1:30 p.m.
or broken axes, small multiples, and reduction of eye movement for easier
decoding of the data. We also examine ways in which information can be
obscured or misrepresented in a graph.
Christopher Battiston, Hospital For Sick Children
Paper 363-2013
The aim of this paper will be an extremely gentle introduction to the very
exciting and somewhat intimidating world of PROC TEMPLATE and Graph
Template Language (GTL). As these two SAS® features are still relatively
new, not many people have had time to learn to learn them and see what
they are capable of accomplishing with minimal effort.
9:30 a.m.
Creating a Useful Operational Report
Andrew Hummel, Delta Air Lines
Robert Goldman, Delta Air Lines
Paper 364-2013
Metrics and reports are highly valued by operational decision makers in
order to make informed and data driven conclusions. However, there is a
balance between presenting useful organized knowledge and displaying
page upon page of raw useless data. SAS® has the power to produce a wide
range of sophisticated graphs that are meaningful. The challenge is to
produce a graph that quickly and accurately measures and displays the
real-world operation while allowing decision makers to make operationally
beneficial determinations. There are numerous SAS papers that give stepby-step instructions on how to build a graph; this is not such a paper. The
goal of this paper is to show how we approached the challenge of building
an operational report and the techniques used.
10:00 a.m.
Cascading Style Sheets: Breaking Out of the Box of ODS
Styles
Kevin Smith, SAS
Paper 365-2013
While CSS has been available in various forms in ODS since SAS® 9.2, SAS 9.4
is the first version that fully utilizes the new style engine’s architecture.
Using the new CSS engine, it is possible to apply custom styles based on
column names, BY group names, BY group variables and anchor names. It is
also possible to specify dynamic attribute values using symget, resolve,
dynamic and expression just like in PROC TEMPLATE styles. If you want to
break out of the box of ODS styles and do some truly original styling, the
CSS techniques in this paper will take you there.
Go Mobile with the ODS EPUB Destination
David Kelley, SAS
Julianna Langston, SAS
Paper 368-2013
The Base SAS® Output Delivery System (ODS) makes it easy to generate
reports for viewing on desktops. What about mobile devices? If you need
on-the-go reports, then the new SAS 9.4 ODS EPUB destination is the ticket.
With ODS EPUB, you can generate your reports as e-books that you can
read with iBooks® on the iPad®, or you can write an e-book from scratch.
This paper provides an introduction to writing e-books with ODS EPUB.
Please bring your iPad, iPhone® or iPod® so that you can download and read
the examples.
2:30 p.m.
Using Design Principles to Make ODS Template
Decisions
Helen Smith, RTI International
Susan Myers, RTI International
Paper 369-2013
With the Output Delivery System (ODS), SAS® continues to provide
programmers with many style templates for developing reports. These
default templates and style definitions present the data in a clear and
attractive manner often with no further thought needed. However, when
producing complicated reports with multiple requirements, using basic
design principles to determine which template or which custom style
definition to use can make for a more readable and comprehensive final
report. This paper presents the code and the design considerations for two
ODS reports; one, a redesign of a 10-plus-year-old SAS program originally
designed with PUT statements, and two, a highly customized SAS program
for delivering output in Microsoft Excel.
3:00 p.m.
Extended SAS® GIFANIM Device Usage on Table
Reporting and Template-Based Graphics
Xin Zhang, Emory University
Neeta Shenvi, Emory University
Azhar Nizam, Emory University
Paper 375-2013
11:00 a.m.
Turn Your Plain Report into a Painted Report Using ODS
Styles
Cynthia Zender, SAS
Allison Booth, SAS
Paper 366-2013
To use STYLE= statement level overrides, you have to understand what
pieces or areas of PRINT, REPORT, and TABULATE output you can change.
And then you have to understand how and where in your procedure syntax
you use the STYLE= override syntax. Last, but not least, you have to know
Dynamic, rather than static, graphs and tables often are more effective,
interactive, and audience-engaging presentations. The SAS® GIFANIM
device enables analysts to create GIF file-based slide shows for web and
PowerPoint presentations, but it only supports device-based graphics and
does not support SG procedure graphics. The GIFANIM device does not
provide an animation of summary tables. In this paper, several ways of
animating stand-alone summary tables, SG procedure graphics, and
graphics with embedded tables, using combinations of the SAS DATA Step
Graphics Interface (DSGI), printer-based methods, and Annotate data sets
are explored. The advantages and disadvantages of each method are
evaluated.
www.sasglobalforum.org/2013
57
3:30 p.m.
Retail — Room 3014
Free Expressions and Other GTL Tips
8:00 a.m.
Prashant Hebbar, SAS
Sanjay Matange, SAS
Paper 371-2013
The Graph Template Language (GTL) provides many powerful features for
creating versatile graphs. The Statistical Graphics Engine in GTL provides
novel ways of using expressions to simplify the task of data preparation.
This presentation covers some new ways of using DATA step functions to
create grouped effect plots based on conditions and to select a subset of
the observations. It also illustrates using PROC FCMP functions in GTL
expressions. Novel uses of non-breaking space for creating small-multiples
graphs and graphs with indented text are discussed. Learn how to express
yourself with ease, graphically!
4:30 p.m.
Analysis and Visualization of E-mail Communication
Using Graph Template Language
Atul Kachare, SAS
Paper 372-2013
Email plays an important role in the corporate world as a means for
collaboration and sharing knowledge. Analyzing email information reveals
useful behavioral patterns that usually carry implicit information regarding
the senders common activities and interests. This paper demonstrates how
we can use SAS® data processing capabilities to extract vital information
based on corporate email communication data by reading email data and
graphically visualizing different communication patterns using Graph
Template Language. This analysis reveals different usage patterns: timebased email volumes with milestones; email exchanges within and across
groups; and communication preferences such as small emails, image-heavy
emails and the words most commonly used in communication.
5:00 p.m.
Visual Techniques for Problem Solving and Debugging
Who Said Change Was Easy
Scott Sanders, Sears
Allan Beaver, Soebeys
Margaret Pelan, Hudson Bay Company
Marty Anderson, Belk
(Invited) Paper 383-2013
In the modern business environment, organizations face rapid change like
never before. Due to the growth of technology, modern organizational
change is largely motivated by exterior innovations rather than internal
moves. When these developments occur, the organizations that adapt
quickest create a competitive advantage for themselves, while the
companies that refuse to change get left behind. Hear how 3 companies
have dealt with Change management and the lessons learned!
9:30 a.m.
SAS® Visual Analytics and Mobile Reporting
Frank Nauta, SAS
Paper 389-2013
Retailers are striving for an omnichannel, customer-centric experience with
brand consistency across all available touch points. It is vital to understand
customers well enough to anticipate their behaviors, know their
preferences and predict when those behaviors and preferences will change.
Data is key to gaining this level of insight. The sheer volume and complexity
of retail data can be a challenge, as can the inability to determine which
variables are relevant to your business. SAS Visual Analytics changes this
with literally a touch of the mouse! This session demonstrates analyzing
hundreds of millions of retail transactions so you can derive insights on ALL
your data in matter of a few seconds and distribute that info to users in an
easy-to-consume manner.
10:30 a.m.
Andrew Ratcliffe, RTSL.eu
Revenue Optimization: How Do You Price?
No matter how well we plan, issues and bugs inevitably occur. Some are
easily solved, but others are far more difficult and complex. This paper
presents a range of largely visual techniques for understanding,
investigating, solving, and monitoring your most difficult problems.
Whether you have an intractable SAS® coding bug or a repeatedly failing
SAS server, this paper offers practical advice and concrete steps to get you
to the bottom of the problem. Tools and techniques discussed in this paper
include Ishikawa (fishbone) diagrams, sequence diagrams, tabular matrices,
and mind maps.
(Invited) Paper 385-2013
(Invited) Paper 373-2013
Brenda Carr, Hudsonâs Bay Company Canada
Hudson’s Bay Company, Markdown Optimization – Our Implementation
and Roll-Out Experience, a case study. To further optimize our markdown
spend and benefit from markdown optimization at a style/store level, HBC
upgraded to SAS® Markdown Optimization 4.3. Selling at full price longer
where we can, while still ensuring we achieve our overall seasonal sell
through target allows HBC to fully maximize their markdown spend and
reap the benefits of increased sales and gross margin in better trending
stores. In this session, you will hear about the path HBC took to roll out this
top initiative, how we gained top-down support of the process, and how
we interacted with our business partners to make this tool one that could
be utilized by all areas.
1:30 p.m.
Retailing in the Era of Tech Titans
Lori Schafer, SAS
Paper 387-2013
Over the past decade, five companies have emerged with the potential to
aggressively reshape the landscape of multiple industries – and to change
marketing as we know it. They are the tech titans: Amazon, Apple, eBay,
58
www.sasglobalforum.org/2013
Facebook and Google. Collectively, these companies are worth more than
$1 trillion. Their growth, cash and vision make them formidable
competitors in any industry and complex partners for any company. These
organizations don’t recognize borders – they are marching beyond the
walls of tech into retailing, advertising, publishing, movies, television,
communications, financial services, and eventually into health care and
insurance. The session highlights the strategies of these companies;
retailers may want to consider their own markets and what may happen
because of the tech titans.
3:00 p.m.
Sobeys and SAS - How Do You Talk to Your Customer?
Ashok Setty, Sobeys Inc.
Wanda Shive, SAS
(Invited) Paper 388-2013
For years, retailers have struggled with measuring the effectiveness of their
promotional advertising efforts. Harnessing the “big data” within their
customer and transaction files continues to be a major challenge.
Approaches for gleaning actionable customer insights from that data are
becoming more common. Measuring total shopping behavior in
conjunction with specific promotion offers provides a better understanding
of the overall impact on profitability. This paper describes how retailers are
utilizing customer analytics to measure the effect that mass promotions
have on the total basket spend of customers and to identify the most
relevant offers for each individual customer.
4:00 p.m.
SAS® Retail Road Map
Saurabh Gupta, SAS
Paper 386-2013
The goal of this presentation is to provide a retail-specific “State of the
Application” update to the SAS® Retail User Group (SRUG) membership. The
session covers the retail solution modules updates released in the past year
and the road map moving forward. This is our forum for the SRUG
membership to hear from the SAS team and vice versa.
5:00 p.m.
Roundtable Discussion: SAS® Integrated Merchandise
Planning
Amy Clouse, Dick's Sporting Goods
(Invited) Paper 390-2013
This session is designed to be a general discussion with the SRUG
membership on the SAS® Integrated Merchandise Planning Solution. We
will provide an opportunity to ask questions and learn how your peers are
gaining the most value from this SAS® solution.
SAS and Big Data — Room 3001
8:00 a.m.
Leveraging Big Data Using SAS® High-Performance
Analytics Server
Priya Sharma, SAS
Paper 399-2013
With the buzz around big data, see how SAS® High-Performance Analytics
Server enables organizations to create greater business value. This paper
aims at providing best practices and techniques when dealing with big data
analytics. Some of the discussed topics are (1) different methods to load
data for Teradata and Greenplum, (2) checking distribution of loaded data
on all nodes, and (3) new HPDS2 procedure in SAS 9.3. Join this discussion
to learn how SAS High-Performance Analytics Server enables you to use 100
percent of your data to get more precise insights and build complex
models at breakthrough speed, and to see results from an example model.
9:00 a.m.
Big Data Meets Text Mining
Zheng Zhao, SAS
Alicia Bieringer, SAS
James Cox, SAS
Russell Albright, SAS
Paper 400-2013
Learning from your customers and your competitors has become a real
possibility because of the massive amount of Web and social media data
available. However, this abundance of data requires significantly more time
and computer memory to perform analytical tasks. This paper introduces
high-performance text mining techniques for SAS® High-Performance
Analytics. Text parsing, text filtering and dimension reduction are
performed using the new HPTMINE procedure in the SAS High-Performance
Analytics Server and are accessed conveniently from within SAS® Enterprise
Miner™. The paper demonstrates and discusses the advantages of this new
SAS functionality and provides real-world examples of the kinds of
performance improvements you can expect.
10:00 a.m.
Uncovering Patterns in Textual Data with SAS® Visual
Analytics and SAS Text Analytics
Mary Osborne, SAS
Justin Plumley, SAS
Dan Zaratsian, SAS
Paper 403-2013
SAS® Visual Analytics is a powerful tool for exploring big data to uncover
patterns and opportunities hidden in your data. The challenge with big
data is that the majority is unstructured data, in the form of customer
feedback, survey responses, social media conversation, blogs and news
articles. By integrating SAS Visual Analytics with SAS Text Analytics, you can
uncover patterns in big data, while enriching and visualizing your data with
customer sentiment and categorical flags, and uncovering root causes that
primarily exist within unstructured data. This paper highlights a case study
that provides greater insight into big data and demonstrates advanced
visualization, while enhancing time to value by leveraging SAS Visual
Analytics high-performance, in-memory technology, Hadoop, and SAS’
advanced text analytics capabilities.
www.sasglobalforum.org/2013
59
11:00 a.m.
9:30 a.m.
SAS® and Hadoop: The BIG Picture
Using SAS® Enterprise Guide®: A System Administrator’s
Perspective
Paul Kent, SAS
Paper 402-2013
SAS® and Hadoop are made for each other. This talk explains some of the
reasons why they are such a good fit. Examples are drawn from the
customer community to illustrate how SAS is a good addition to your
Hadoop cluster.
SAS® Enterprise Guide® Implementation and
Usage — Room 3002
8:00 a.m.
Consistent and Organized Analysis: Moving Beyond Piein-the-Sky to Actual Implementation via SAS® Enterprise
Guide®
Amber Schmitz, Prime Therapeutics
Paper 404-2013
Consistency and timeliness are two goals that every reporting department
strives to achieve. SAS® Enterprise Guide® provides tools that support these
goals that many analysts overlook. The goal of this paper is to demonstrate
the use of SAS Enterprise Guide to construct a project template that can be
used to report standard metrics for various clients. The template is built
around SAS Enterprise Guide tools that allow for consistent analysis
methods, project organization, and version control documentation. We
explore: 1) Utilizing program code for efficient data pulls, 2) Using SAS
Enterprise Guide tasks for analysis, 3) Exploiting built-in tools such as Notes
and Process Flows for template organization and version control, 4)
Demonstrating business intelligence via built-in graphs.
8:30 a.m.
Haibo Jiang, Allergan, Inc.
Paper 406-2013
This paper shares our experience of supporting SAS® Intelligence Platform
server and client products with other Platform Administrators. We will focus
on SAS® Enterprise Guide® as a client application on Windows desktop, and
SAS servers (SAS Metadata Server®, Object Spawner, and SAS Application
Servers) installed on Hewlett-Packard (HP) UNIX machine. The content of
this paper will describe technical details related to user- and system-related
activities in the following areas: [] Starting SAS servers and checking their
status. [] Connection to SAS servers, user authorization, and authentication
[] Initiation of requests from SAS Enterprise Guide, and using SAS
Workspace Server and Stored Process Server. [] Performance considerations
for the processing and presentation of analysis results in SAS Enterprise
Guide.
10:00 a.m.
Statistical Analyses Using SAS® Enterprise Guide®
Scott Leslie, MedImpact Healthcaree Systems, Inc.
(Invited) Paper 407-2013
Conducting statistical analyses involves choosing proper methods,
understanding model assumptions and displaying clear results. The latest
releases of SAS® Enterprise Guide® offer conveniences, such as point-andclick wizards and integrated syntax help, to ease the burden on users. This
tutorial demonstrates how to perform statistics quickly and easily using
some handy features of SAS Enterprise Guide. Examples of multiple linear
regression, logistic regression, and survival analysis are covered as well as
some hints on how to navigate SAS Enterprise Guide menus. This tutorial is
intended for SAS® users with beginning to intermediate experience with
the above-mentioned statistics or those with little SAS Enterprise Guide
experience.
The Concepts and Practice of Analysis with SAS®
Enterprise Guide®
11:00 a.m.
(Invited) Paper 405-2013
Aaron Hill, MDRC
Chris Schacherer, Clinical Data Management Systems, LLC
Due in part to its success helping SAS® programmers leverage their
development skills against the challenges of creating analytic solutions in a
new environment, SAS® Enterprise Guide® continues to gain acceptance as
an enterprise solution for reporting and analytic applications. For
organizations to realize maximum benefit from their investment in SAS
Enterprise Guide, subject-matter experts and a new generation of "SAS
naive" analysts also need to be trained in the use of this tool. The current
work provides a framework for this training by explaining the relationship
between SAS Enterprise Guide and traditional SAS programming,
introducing the basic SAS Enterprise Guide concepts necessary to function
in this environment, and presenting examples of common tasks that will
help these users become immediately productive.
60
www.sasglobalforum.org/2013
Destination Known: Programmatically Controlling Your
Output in SAS® Enterprise Guide®
Paper 408-2013
In a SAS® Enterprise Guide® project with multiple reports and graphics, you
can organize output by selectively sending content to different ODS
destinations embedded within the project. For example, within a single
program, you can embed tables in HTML format, graphics in a SAS® report,
and other output in text. The SAS syntax is simple and gives you
programmatic control over all output and destinations. The result: a wellorganized project with all results in their preferred format.
11:30 a.m.
2:30 p.m.
A tour of new features in SAS Enterprise Guide 4.3, 5.1,
and 6.1
For All the Hats You Wear: SAS® Enterprise Guide® Has
Got You Covered
Lina Clover, SAS
Anand Chitale, SAS
I-kong Fu, SAS
Paper 409-2013
As an Enterprise Guide user or administrator, you have probably recently
upgraded to version 4.3 or 5.1 or are looking forward to going to these
versions or the new 6.1 version. In this paper, we will take you on a tour of
new features available to you in each of these releases so that you can be
more productive with your current or upcoming version as well as obtain a
preview of what's coming in the future. We will cover the differences
between the various versions of SAS Enterprise Guide clients and their
compatibility with the corresponding SAS Server versions, and explain how
to upgrade from your current version with the goal of giving you better
decision points for planning your client /server upgrade strategy.
1:30 p.m.
SAS® Enterprise Guide®: More Than a Gift from Outer
Space
Tricia Aanderud, And Data Inc
Paper 410-2013
SAS® Enterprise Guide® seem alien to you? Let's walk through the many SAS
Enterprise Guide features using some UFO sightings data. During the
presentation, you will learn some basics, how to change the advanced
options, and also explore some newer features of SAS Enterprise Guide.
Whether a SAS® programmer or an experienced SAS Enterprise Guide user,
you will leave with some practical tips and learn what sightings were
reported to the UFO websites.
2:00 p.m.
Using VBA to Debug and Run SAS® Programs
Interactively, Run Batch Jobs, Automate Output, and
Build Applications
FENG LIU, Genworth Financial
Ruiwen Zhang, SAS
Paper 411-2013
SAS® Enterprise Guide® provides an API that lets users automate almost
every aspect of running Enterprise Guide projects or even SAS® programs.
Visual Basic for Applications (VBA) under Excel provides a rich environment
for debugging and running VBA applications. This paper shows how to use
VBA to access the automation API of Enterprise Guide to do sophisticated
tasks or build your own applications. VBA lets you create SAS programs on
the fly, debug and run programs, analyze SAS “lists,” write logs to files, and
examine SAS ODS. You can accomplish tasks like running PROC EXPORT
automatically, which is not feasible through Enterprise Guide’s main
interface. Advanced SAS users can run batch jobs, schedule jobs in parallel,
or use SAS output as input to other applications.
Chris Hemedinger, SAS
Paper 412-2013
Are you new to SAS® and trying to figure out where to begin? Are you a SAS
programmer, already comfortable with code but unsure about new tools?
Are you a statistician seeking to apply your techniques in a new way? Are
you a data manager, just trying to get your data in shape? Perhaps you're a
Jack (or Jill) of all trades trying to manage work in the simplest way
possible. Regardless of your background, SAS® Enterprise Guide® is full of
features and techniques to help you achieve your objective. In this paper,
we show how you can turn SAS Enterprise Guide into your tool to get work
done immediately, without conforming to an entirely new way of working
just to become productive.
3:30 p.m.
Finally, a Tool for Business Users! A Step-By-Step
Practical Approach to Pharma Sales Reporting Using
SAS® Enterprise Guide® 4.3
Ramya Purushothaman, Cognizant Technology Solutions
Airaha Chelvakkanthan Manickam, Cognizant Technology
Solutions
Paper 413-2013
SAS® Enterprise Guide® can be considered the integrated development
environment (IDE) for SAS® users. SAS Enterprise Guide has powerful data
management capabilities, a sophisticated Query Builder, and data
sampling, ranking, transposing, and even creating and editing data
capabilities. This paper presents a real-time case study of reporting Pharma
sales using SAS Enterprise Guide. It includes capturing drug sales dollars
data and performing business transformations, summarization, and
producing summary charts for executive reports and dashboards. The goal
of this paper is to educate SAS users on how all of these actions are easily
performed by a series of simple clicks in SAS Enterprise Guide 4.3.
4:00 p.m.
Stealing the Admin's Thunder: SAS® Macros to Interact
with the UNIX OS from within SAS® Enterprise Guide®
Thomas Kunselman, Southern California Edison
Paper 414-2013
For SAS® users who are unfamiliar with the UNIX environment, simple tasks
like copying, renaming, or changing the permission settings on a file can be
very non-intuitive. Many of these tasks are not even possible through the
SAS® Enterprise Guide® Server List Window. This paper will present several
SAS macros that can be used to: view and kill UNIX host processes; display,
compare, and manage folders and files, including copying subfolders and
changing permissions and owners; display and set default file system
permissions for new objects. Please note that for these macros to work, the
X command must be allowed on the SAS server.
www.sasglobalforum.org/2013
61
4:30 p.m.
Statistics and Data Analysis — Room 3016
Improving Your Relationship with SAS® Enterprise
Guide: Tips from SAS® Technical Support
9:30 a.m.
Jennifer Bjurstrom, SAS
Paper 415-2013
SAS® Enterprise Guide® has proven to be a very beneficial tool for both
novice and experienced SAS® users. Because it is such a powerful tool, SAS
Enterprise Guide has risen in popularity over the years. As a result, SAS
Technical Support consultants field many calls from users who want to
know the best way to use the application to accomplish a task or to obtain
the results they want. This paper encompasses many of the tips that SAS
Technical Support has provided to customers over the years. These tips are
designed to improve your proficiency with SAS Enterprise Guide in the
areas of connection and configuration, workflow preferences, logging, data
manipulation, project files, X commands, and custom tasks.
Statistics and Data Analysis — Room 3016
8:00 a.m.
Finding the Gold in Your Data: An Overview of Data
Mining
David Dickey, NC State University
(Invited) Paper 501-2013
"Data mining" has appeared often recently in analytic literature and even in
popular literature, so what exactly is data mining and what does SAS®
provide in terms of data mining capabilities? The answer is that data mining
is a collection of tools designed to discover useful structure in large data
sets. With an emphasis on examples, this talk gives an overview of methods
available in SAS® Enterprise Miner™ and should be accessible to a general
audience. Topics include predictive modeling, decision trees, association
analysis, incorporation of profits, and neural networks. We'll see that some
basic ideas underlying these techniques are related to standard statistical
techniques that have been around for some time but now have been
automated to become more user friendly.
Statistics and Data Analysis — Room 2005
9:30 a.m.
Current Directions in SAS/STAT® Software Development
Maura Stokes, SAS
Paper 432-2013
Recent years have brought you SAS/STAT® releases in rapid succession, and
another one is coming in 2013. Which new software features will make a
difference in your work? What new statistical trends should you know
about? This paper describes recent areas of development focus, such as
Bayesian analysis, missing data methods, postfitting inference, quantile
modeling, finite mixture models, specialized survival analysis, structural
equation modeling, and spatial statistics. The paper introduces you to the
concepts and illustrates them with practical applications.
From Big Data to Big Statistics
John Sall, SAS
Paper 434-2013
Now that we have lots of data and can process it amazingly fast, we still
need ways to look at it without being overwhelmed. We don’t want to look
at 10,000 graphs--we want one graph that shows the bright spots among
10,000 graphs. We need volcano plots and false-discovery-rate plots. We
want the computer and software to do the work of finding what is most
interesting and bringing it to our attention. We want our results sorted and
summarized, but with access to the detail we need to understand it. Also,
when we look at the most significant of thousands of statistical tests, we
want to know if we are seeing random coincidence selected out of
thousands, or if we are seeing real effects.
Statistics and Data Analysis — Room 2005
10:30 a.m.
DICHOTOMIZED_D: A SAS® Macro for Computing Effect
Sizes for Artificially Dichotomized Variables
Patrice Rasmussen, 5336 Clover Mist Drive
Isaac Li, Univ. of South Florida
Patricia Rodriguez de Gil, University of South Florida
Jeanine Romano, University of South Florida
Aarti Bellara, University of South Florida
Harold Holmes, University of South Florida
Yi-Hsin Chen, University of South Florida
Jeffrey Kromrey, University of South Florida
Paper 491-2013
Measures of effect size are recommended to communicate information
about the strength of relationships between variables, providing
information to supplement the reject/fail-to-reject decision obtained in
statistical hypothesis testing. With artificially dichotomized response
variables, seven methods have been proposed to estimate the standardized
mean difference effect size that would have been realized before
dichotomization. This paper provides a SAS® macro, DICHOTOMIZED_D, for
computing these seven effect size estimates by utilizing data from FREQ
procedure output data sets. The paper provides the macro programming
language, as well as results from an executed example of the macro.
Statistics and Data Analysis — Room 3016
10:30 a.m.
A Multilevel Model Primer Using SAS® PROC MIXED
Bethany Bell, University of South Carolina
Mihaela Ene, University of South Carolina
Whitney Smiley, University of South Carolina
Jason Schoeneberger, University of South Carolina
Paper 433-2013
This paper provides an introduction to specifying multilevel models using
PROC MIXED. After a brief introduction to the field of multilevel modeling,
users are provided with concrete examples of how PROC MIXED can be
used to estimate (a) two-level organizational models, (b) two-level growth
62
www.sasglobalforum.org/2013
models, (c) three-level organizational models, and (4) three-level growth
models. Both random intercept and random intercept and slope models are
illustrated. Examples are shown using Early Childhood Longitudinal Study–
Kindergarten cohort data. For each example, narrative explanations are
accompanied by annotated examples of the PROC MIXED code and
corresponding output. Users are also introduced to examining model fit
using the SAS® macro MIXED_FIT as well as checking the distributional
assumptions for two-level models using the SAS macro MIXED_DX.
Statistics and Data Analysis — Room 2005
11:00 a.m.
The Value of Neighborhood Information in Prospect
Selection Models: Investigating the Optimal Level of
Granularity
Philippe Baecke, Vlerick Business School
Dirk Van den Poel, Ghent University
(Invited) Paper 492-2013
Within analytical customer relationship management (CRM), customer
acquisition models suffer the most from a lack of data quality because the
information of potential customers is mostly limited to socio-demographic
and lifestyle variables obtained from external data vendors. Particularly in
this situation, taking advantage of the spatial correlation between
customers can improve the predictive performance of these models. This
study compares the predictive performance of an autoregressive and
hierarchical technique in an application that identifies potential new
customers for 25 products and brands. In addition, this study shows that
the predictive improvement can vary significantly depending on the
granularity level on which the neighborhoods are composed. Therefore, a
model is introduced that simultaneously incorporates multiple levels of
granularity resulting in even more accurate predictions.
Statistics and Data Analysis — Room 2005
1:30 p.m.
Having an EFFECT: More General Linear Modeling and
Analysis with the New EFFECT Statement in SAS/STAT®
Software
Phil Gibbs, SAS
Randy Tobias, SAS
Kathleen Kiernan, SAS
Jill Tao, SAS
Paper 437-2013
Linear models relate a response to a linear function of a design matrix X.
General linear models, long available in standard SAS/STAT® 9.3 procedures
such as GLM and MIXED, incorporate classification, interaction, and
crossproduct effects to define X. The new EFFECT statement, which is
available in many SAS/STAT 9.3 procedures, extends how you can define X.
It enables you to fit models with nonparametric regression effects,
crossover and carryover effects, and complicated inheritance effects. This
paper first shows how the EFFECT statement fits into the general
architecture of SAS/STAT linear modeling tools, and then explains and
demonstrates specific effect types. You will see how this powerful new
feature easily enhances the statistical analyses that you can perform.
Statistics and Data Analysis — Room 2007
1:30 p.m.
Missing No More: Using the MCMC Procedure to Model
Missing Data
Fang Chen, SAS
Paper 436-2013
Statistics and Data Analysis — Room 3016
11:00 a.m.
Joint Modeling of Mixed Outcomes in Health Services
Research
Joseph Gardiner, Michigan State University
(Invited) Paper 435-2013
Outcomes with different attributes, of continuous, count, and categorical
types, are often encountered jointly in many settings. For example, two
widely used measures of healthcare utilization, length of stay (LOS) and
cost, can be analyzed jointly with LOS as a count and cost as continuous.
Occurrence of an adverse event (binary) would impact both outcomes. For
fitting marginal distributions and assessing the impact of explanatory
variables on outcome, SAS offers a number of procedures. Correlation and
clustering are additional features of these outcomes that must be
addressed in analyses. This paper surveys the GLIMMIX, COPULA, PHREG,
and QLIM procedures, which can be applied to modeling multivariate
outcomes of mixed types. Examples from the literature are used to
demonstrate the application of these procedures.
Missing data are often a problem in statistical modeling. The Bayesian
paradigm offers a natural model-based solution for this problem by
treating missing values as random variables and estimating their posterior
distributions. This paper reviews the Bayesian approach and describes how
the MCMC procedure implements it. Beginning with SAS/STAT® 12.1, PROC
MCMC automatically samples all missing values and incorporates them in
the Markov chain for the parameters. You can use PROC MCMC to handle
various types of missing data, including data that are missing at random
(MAR) and missing not at random (MNAR). PROC MCMC can also perform
joint modeling of missing responses and covariates.
Statistics and Data Analysis — Room 2005
2:30 p.m.
Regression of NASCAR: Looking into Five Years of
Jimmie Johnson
Yun Gao, California State Universiday Long Beach
Paper 439-2013
In this paper, we investigate the winnings associated with different factors
for NASCAR drivers. We want to predict the winnings that a driver can earn
in a season given other, related factors, such as the number of races the
driver competes in, the average finish position, or the make of car. We
obtained 190 observations with 15 factors and randomly split the data into
learning data and test data. Using the learning data set, we conducted
multiple regression analyses to build a predictive model. Then we
www.sasglobalforum.org/2013
63
examined the final model with the test data set to see how well the model
would work in the future. The model shows a high degree of accuracy in
predicting the future.
Statistics and Data Analysis — Room 2007
variance estimates. Likelihood ratio testing is a more flexible approach, as it
can be used to compare models that differ in both fixed and random
effects. The likelihood ratio test statistic requires a complex calculation that
is not included in PROC MIANALYZE. This paper describes a SAS macro,
MMI_ANALYZE, that fits two user-specified models in PROC MIXED, pools
the estimates from those models (including variance components), and
implements a pooled likelihood ratio test.
2:30 p.m.
A SAS® Macro for Applying Multiple Imputation to
Multilevel Data
Statistics and Data Analysis — Room 2005
Paper 438-2013
Multilevel Reweighted Regression Models to Estimate
County-Level Racial Health Disparities Using PROC
GLIMMIX
Stephen Mistler, Arizona State University
Single-level multiple imputation procedures (e.g., PROC MI) are not
appropriate for multilevel data sets where observations are nested within
clusters. Analyzing multilevel data imputed with a single-level procedure
yields variance estimates that are biased toward zero and may yield other
biased parameters. Given the prevalence of clustered data (e.g., children
within schools; employees within companies; observations within people),
a general approach is needed for handling missing data in multilevel data
sets. This paper describes a SAS® macro, MMI_IMPUTE, that performs
multiple imputation for clustered data sets with two levels. The macro uses
a Bayesian implementation of the mixed linear model to generate
imputations for lower-level incomplete variables, and uses single-level
procedures similar to those used in PROC MI to generate imputations for
cluster-level variables.
Statistics and Data Analysis — Room 2005
3:00 p.m.
Short-Term Costs of Smoking during Pregnancy: A
Geometric Multidimensional Approach
3:30 p.m.
Melody S. Goodman, Division of Public Health Sciences at
Washington University in St. Louis School of Medicine
Lucy D'Agostino, Division of Public Health Sciences,
Department of Surgery, Washington University School of
Medicine
Paper 442-2013
The agenda to reduce racial health disparities has been set primarily at the
national and state levels. These levels may be too far removed from the
individual level where health outcomes are realized, and this disconnect
may be slowing the progress made in reducing these disparities. This paper
focuses on establishing county-level prevalence estimates of diabetes
among non-Hispanic whites and non-Hispanic blacks. These estimates use
multilevel reweighted regression models through the GLIMMIX procedure
with 2010 Behavioral Risk Factor Surveillance System data and 2010 census
data. To examine whether racial disparities exist at the county level, the
paper estimates the risk difference of prevalence estimates between races.
It subsequently ranks counties and states by the magnitude of disparities.
Violeta Balinskaite, University of Bologna
Paper 441-2013
Smoking during pregnancy imposes a considerable economic burden on
society. This phenomenon has been studied fairly extensively in the United
States, but little is known about its costs within the European Union. This
paper attempts to estimate the additional neonatal costs of a mother in the
European Union who smokes during pregnancy compared to the
alternative of her not smoking. The geometric multidimensional approach
that is used for analysis involves the use of conditional multiple
correspondence analysis as a tool for investigating the dependence
relationship between covariates and the assignment-to-treatment indicator
variable within a strategy whose final aim is to find balanced groups.
Statistics and Data Analysis — Room 2007
3:00 p.m.
A SAS® Macro for Computing Pooled Likelihood Ratio
Tests with Multiply Imputed Data
Stephen Mistler, Arizona State University
Paper 440-2013
For multilevel analyses (e.g., linear mixed models), researchers are often
interested in pooling, interpreting, and testing both fixed effects and
random effects. PROC MIANALYZE has two shortcomings in this regard.
First, it cannot easily pool variance estimates. Second, the significance tests
of these estimates are Wald-type tests that are inappropriate for testing
64
www.sasglobalforum.org/2013
Statistics and Data Analysis — Room 2007
3:30 p.m.
Commuting Time and Accessibility in a Joint Residential
Location, Workplace, and Job Type Choice Model
Ignacio A. Inoa, Université de Cergy-Pontoise
Paper 443-2013
The effect of an individual-specific measure of accessibility to jobs is
analyzed using a three-level nested logit model of residential location,
workplace, and job-type choice. This measure takes into account the
attractiveness of different job types when the workplace choice is
anticipated in the residential location decision. The model allows for
variation in the preferences for job types across individuals and accounts
for individual heterogeneity of preferences at each choice level in the
following dimensions: education, age, gender, and children. Using data
from the Greater Paris Area, estimation results indicate that the individualspecific accessibility measure is an important determinant of the residential
location choice and its effects strongly differ along the life cycle.
Statistics and Data Analysis — Room 2005
4:00 p.m.
Models for Ordinal Response Data
Robin High, University of Nebraska Medical Center
Paper 445-2013
The types of computations with response data having ordered categories
with SAS® procedures are not well known. Various models can be evaluated
through programming statements entered into PROC NLMIXED including
the partial proportional odds, adjacent logit, continuation ratio, and
stereotype models. The process requires no restructuring of the input data
set, as required with procedures that can produce a few of these models.
The correct interpretation of ordinal logistic regression models depends on
how both the response and explanatory data are coded and if any formats
are applied. Implementation of these models assumes a background with
general linear models and categorical data analysis including maximum
likelihood equations and computing odds ratios with binary data.
Statistics and Data Analysis — Room 2007
4:00 p.m.
Examining Mediator and Indirect Effects of Loneliness in
Social Support on Social Well-Being Using the Baron and
Kenny Method and a Bootstrapping Method
Abbas Tavakoli, University of South Carolina
Sue Heiney, University of South Carolina
Paper 444-2013
This study examines the mediator effect and the indirect effect of loneliness
in social support on social well-being by using two methods: the Baron and
Kenny method and a bootstrapping method. The cross-sectional data come
from a longitudinal randomized trial design that had 185 participants.
Baron and Kenny steps and Hayes were used to examine the mediator
effect. The Baron and Kenny results indicate no mediator effect for
loneliness in the relationship between social support and social well-being.
Bootstrapping results indicate that the direct effect was 0.591 (95% CI:
0.589-0.593 for normal theory and 0.481- 0.690 for percentile) and the
indirect effect was 0.040 (95% CI: 0.039-0.040 for normal theory and
0.006-0.087 for percentile). The results show that both methods have
significant indirect effect.
Statistics and Data Analysis — Room 2005
4:30 p.m.
Ordinal Response Modeling with the LOGISTIC
Procedure
responses. This paper also discusses methods of determining which
covariates have proportional odds. The reader is assumed to be familiar
with using PROC LOGISTIC for binary logistic regression.
Statistics and Data Analysis — Room 2007
4:30 p.m.
Segmentation and Classification Analysis Using SAS®
Rachel Poulsen, TiVo
(Invited) Paper 447-2013
An idiom in the customer service industry is “the customer is always right”.
However, in many instances the customer will not speak up and another
popular idiom must be used “Actions speak louder than words”. Customer
actions can be measured to infer what they will not say. Once measured,
segmentation analysis can be used to make sense of the large amount of
behavioral data by placing customers into various segments. Classification
models are then used to assign new customers to a segment. Statistical
algorithms used to segment and classify observations include Collaborative
Filtering and Machine Learning Models. This paper will illustrate how SAS®
can be used to segment and classify observations using the FASTCLUS and
DISCRIM procedures.
Systems Architecture and Administration — Room
2006
8:00 a.m.
Enhance Your High Availability Story by Clustering Your
SAS® Metadata Server in SAS® 9.4
Bryan Wolfe, SAS
Amy Peters, SAS
Paper 468-2013
It can be challenging to use clustered file systems, backups and system
tools to ensure SAS® Metadata Server is always available. In SAS 9.4, you can
create and manage a clustered metadata deployment using SAS tools to
remove single-point-of-failure concerns for the SAS Metadata Server. The
clustered SAS Metadata Server keeps its data in sync, balances its load and
continues to handle requests if a node fails, all while presenting a single
face to the outside world. Once it is set up, you don’t need to treat the SAS
Metadata Server any differently than you do the single server.
8:30 a.m.
Best Practices for Deploying Your SAS® Applications in a
High-Availability Cluster
Helen Pan, SAS
Bob Derr, SAS
Paper 469-2013
Logistic regression is most often used to model simple binary response
data. Two modifications extend it to ordinal responses that have more than
two levels: using multiple response functions to model the ordered
behavior, and considering whether covariates have common slopes across
response functions. This paper describes how you can use the LOGISTIC
procedure to model ordinal responses. Before SAS/STAT® 12.1, you could
use cumulative logit response functions with proportional odds. In SAS/
STAT 12.1, you can fit partial proportional odds models to ordinal
Are you frustrated when software or hardware failures interrupt your SAS®
system? A high-availability (HA) cluster with HA software can help you by
providing failover protection for your SAS applications, thus reducing your
system downtime. This paper discusses what you should consider when
deploying your SAS applications or servers in an HA cluster, as well as the
following best practices of the deployment process: - HA cluster
architecture with a SAS deployment. - SAS installation and deployment in
an HA cluster. - Dependency configuration of the SAS servers. - Impact on
the SAS clients when a failover happens.
Paper 446-2013
www.sasglobalforum.org/2013
65
9:00 a.m.
11:00 a.m.
The Top Four User-Requested Grid Features Delivered
with SAS® Grid Manager 9.4
Best Practices for Deploying SAS® on Red Hat Enterprise
Linux
Paper 470-2013
The number of SAS deployments on Red Hat Enterprise Linux (RHEL)
continues to increase in recent years because more and more customers
have found RHEL to be the best price/performance choice for new and/or
updated SAS deployments on x86 systems. Back for the third year at SGF,
Shak and Barry will share new performance findings and best practices for
deploying SAS on Red Hat Enterprise Linux and will discuss topics such as
virtualization, GFS2 shared file system, SAS Grid Manager and more. This
session will be beneficial for SAS customers interested in deploying on Red
Hat Enterprise Linux, or existing SAS-on-RHEL customers who want to get
more out of their deployments.
Doug Haigh, SAS
As more and more SAS customers choose to deploy their platform for SAS®
Business Analytics with SAS Grid Manager, we continue to expand the
capabilities of a shared, managed, and highly available environment.
Several features have been added in SAS 9.4, providing the following: easier administration of the different types of SAS users in the grid increased management of more SAS components running in the grid integration of the grid with IT standards, such as an existing enterprise
scheduler - increased debugging capabilities for quick problem resolution
This paper details how to accomplish all of the above through new grid
options sets, workspace servers spawned on the grid, new options in
SASGSUB, and enhanced error logging.
9:30 a.m.
Creating metadata environment from existing one for
testing purpose
Jouni Javanainen, Aureolis
Paper 471-2013
Most organizations have a need for the development, test and production
environments, either in the same physical platform or on separate
platforms. Separate environments may not use the same ports of
communication. Using package of migration from existing metadata, it is
possible to define specific communication ports, so that it does not disturb
the other environments. In addition to this it is needed very little extra
finishing steps, such as paths of directories or libraries. It is important that
this type of environment can be created through a formal process quickly,
reliably and efficiently. SAS has a good set of tools to create this a welldocumented method for creating environments. This paper covers how
easily you can create identical environments for development and testing
purposes.
10:00 a.m.
Configure Your Foundation SAS® Client Easily and
without Risk
Peter Crawford, Crawford Software Consultancy Limited
(Invited) Paper 472-2013
For your client install of Base SAS®, don't use the default provided. Instead,
use the simple features in this paper and presentation to support the
flexibility you want. This method of applying options as SAS starts,
eliminates risky techniques of the past when developers would update the
configuration file provided by SAS. The technique is described and
demonstrated with examples for the Microsoft Windows environment for
Base SAS, but the issues are very similar if launching SAS clients and servers
on UNIX (including UNIX on z/OS). There is a small overhead with this
proposal, but if you review the paper, I’ m sure you will consider it worthy
enough to give it a try.
66
www.sasglobalforum.org/2013
(Invited) Paper 486-2013
1:30 p.m.
SAS Administration: The State of the Art (Panel
Discussion)
Greg Nelson, ThotWave Technologies, LLC.
Michael Raithel, Westat
Paul Homes, Metacoda
Jennifer Parks, CSC Inc
(Invited) Paper 473-2013
The implementation of SAS can take on many forms in organizations
around the globe - from single-server SAS Foundation installs to multi-tenet
SAS solutions. Similarly the role of the SAS administrator has evolved
significantly - especially since the introduction of SAS 9. Join us for this
panel discussion for an in-depth conversation on SAS Administration.
Topics covered will include:
• SAS administrator roles and responsibilities
• Best practices in data management, governance, architecture and
business process integration
• Adoption of new technologies including upgrades and maintenance
• Backup, Recovery, Disaster Recovery
• Multi-tenet architectures
• Optimizing SAS. Panelists will include experts in capacity planning, SAS
metadata, security, optimizing operating environments, system design
and architecture. Panelists will include personnel from SAS, system
integrators and various industry organizations
2:30 p.m.
A Case Study of Tuning an Enterprise Business
Intelligence Application in a Multi-OS Environment
Fred Forst, SAS
Paper 474-2013
A corporation is planning to expand its use of SAS® Web Report Studio by
increasing its user base. Two questions emerge: 1) What happens to SAS
Web Report Studio response time as users are added? 2) What tuning
modifications can be used to improve performance? This case study uses a
SAS Web Report Studio simulation technique to expand the workload and
examine software response times. Logs from SAS Web Report Studio,
WebSphere, NMON and the server tier are parsed and stored in a SAS data
set for all analyses. Many stress test runs are executed, modifying various
tuning parameters. The analysis shows the effects of tuning and offers
insight on best practices.
3:00 p.m.
Grand Designs: Why It Pays to Think About Technical
Architecture Design Before You Act
Simon Williams, SAS
Paper 475-2013
The sustainability of a business is not a short-term goal; it depends on
effective planning and considered thinking. The same principle of
sustainability applies to IT systems, including those using technologies
from SAS. Getting the best out of SAS® technology is easy once all the IT
and business requirements and constraints are clearly defined, understood
and agreed upon as being testable. This paper describes the effective
thinking and systematic approach that can be applied to creating a
sustainable SAS solution. The paper also outlines the advantages to this
approach and describes a simple checklist that can be helpful when
designing a new SAS environment.
3:30 p.m.
Kerberos and SAS® 9.4: A Three-Headed Solution for
Authentication
Stuart Rogers, SAS
Paper 476-2013
Kerberos is a network authentication protocol designed to provide strong
authentication for client/server applications by using secret-key
cryptography. With the release of SAS® 9.4, there are three ways Kerberos
can be used with the SAS platform for business analytics. Kerberos provides
Integrated Windows authentication from a range of clients to a range of
servers. This paper reviews how Kerberos is used with the SAS 9.4 platform
for business analytics. It explores the considerations and constraints when
using Kerberos and summarizes solutions for some common issues.
investigated and implementation client virtualization to simplify the
configuration, deployment and support of their end-user computers. Client
virtualization looks at changing the 1 end-user to 1 desktop computer
paradigm for SAS software installation and finding ways of reducing the
administrative burden associated with the one end user desktop computer
while gaining operational efficiencies and a more robust deployment
model. This paper will focus on the following topics: Primary drivers for
adopting this technology; Census SAS support model; Client virtualization
architecture; Deployment best practices.
5:00 p.m.
SAS® Enterprise Business Intelligence Deployment
Projects in the Federal Sector: Best Practices
Jennifer Parks, CSC Inc
(Invited) Paper 478-2013
Systems engineering life cycles (SELC) in the federal sector embody a high
level of complexity due to legislative mandates, agency policies, and
contract specifications layered over industry best practices, all of which
must be taken into consideration when designing and deploying a system
release. Additional complexity stems from the unique nature of ad-hoc
predictive analytic systems that are at odds with traditional, unidirectional
federal production software deployments to which many federal sector
project managers have grown accustomed. This paper offers a high-level
roadmap for successful SAS® EBI design and deployment projects within
the federal sector. It's addressed primarily to project managers and SAS
administrators engaged in the SELC process for a SAS EBI system release.
4:00 p.m.
SAS® and the New Virtual Storage Systems
Tony Brown, SAS
Margaret Crevar, SAS
Paper 487-2013
Storage providers are offering a wave of new, advanced storage
subsystems. These offerings promise virtualized, thin-provisioned, tiered,
intelligent storage that is easy to manage and will reduce costs. For many
random and mixed workload applications, the promises deliver well and
with good performance. Unfortunately, the SAS® I/O workload profiles tend
to violate some of the primary design assumptions underlying the
configuration of these new systems. This paper will address the SAS
workload-specific issues that need to be considered when configuring
these new storage systems for expected performance.
4:30 p.m.
SAS® Virtual Desktop Deployment at the U.S. Bureau of
the Census
Lori Guido, US Census Bureau
Michael Bretz, SAS
Stephen Moore, US Census Bureau
Paper 490-2013
The U.S. Census Bureau has a SAS® user base of approximately 2600 users
requiring deployment of many SAS client solutions on individual desktops.
Using new deployment strategies, we reduced deployment delivery time
while increasing installation quality and standardization. Census
www.sasglobalforum.org/2013
67
68
www.sasglobalforum.org/2013
Beyond the Basics — Room 2016
11:30 a.m.
9:00 a.m.
A First Look at the ODS Destination for PowerPoint
Creating Graph Collections with Consistent Colors Using
ODS Graphics?
Paper 041-2013
Philip Holland, Holland Numerics Ltd
(Invited) Paper 038-2013
Collections of line graphs or bar charts, where the graph data is grouped by
the same value, are frequently used to identify differences and similarities
in behavior. Unfortunately, by default, the colors used for each line can
change across the graph collection if some group values are not present in
every graph. In SAS/GRAPH®, this problem has been solved by generating
SYMBOL or PATTERN statements based on the data, or using annotation to
create all of the graph lines, bars and legends. Neither of these solutions is
readily available in ODS Graphics. This paper will solve this problem using
macros with PROC SGPLOT and PROC TEMPLATE, giving the user complete
control over how every graph looks.
Beyond the Basics — Room 3016
10:00 a.m.
Renovating Your SAS® 9.3 ODS Output: Tools for
Everything from Minor Remodeling to Extreme
Makeovers
Bari Lawhorn, SAS
Paper 039-2013
The SAS® 9.3 HTML output that is generated with the ODS HTML statement
and the new HTMLBLUE style looks great with the default settings. But
realistically, there are always pieces of your output that you want to
change. Sometimes you just need a little remodeling to get the results you
want; other times, the desired changes call for an extreme makeover. This
paper shows how you can use certain ODS statements (for example, ODS
SELECT and ODS TEXT) and ODS statement options (for example, STYLE=)
for minor remodeling. The paper also illustrates how to use the REGISTRY,
TEMPLATE, and DOCUMENT procedures for a more extreme makeover.
These makeovers apply to HTML as well as other ODS destinations.
Beyond the Basics — Room 2016
10:30 a.m.
"How Do I ...?" There Is More Than One Way to Solve That
Problem; Why Continuing to Learn Is So Important
Art Carpenter, CA Occidental Consultants
(Invited) Paper 029-2013
In the SAS® forums, questions are often posted that start with "How do
I . . . ?". Generally, there are multiple solutions to the posted problem, and
these vary from simple to complex. All too often, the simple solution is both
inefficient and reflects a naive understanding of the SAS language. This
would not be so very bad except sometimes the responder thinks that their
response is the best solution or, perhaps worst, the only solution.
Tim Hunter, SAS
The inclusion of PowerPoint is part of the next generation of ODS
destinations. You can use this destination to send PROC OUTPUT directly
into native PowerPoint format. See examples of slides created by ODS.
Learn how to create presentations using ODS; how to use ODS style
templates to customize the look of your presentations; and how to use
predefined layouts to make title slides and two-column slides. Learn how
the ODS destination for PowerPoint is similar – and different – to other ODS
destinations. Stop cutting and pasting; let the ODS destination for
PowerPoint do the work for you!
12:30 p.m.
Can You Create Another PowerPoint for Me? How to Use
Base SAS® and DDE to Automate Snappy PowerPoint
Presentations
Scott Koval, Pinnacle Solutions, Inc.
Mitchell Weiss, Maguire Associates
Paper 042-2013
Your supervisor appreciates your wonderful and informative SAS® reports.
How many times have you heard, “Great! Now, can you compile all the SAS
reports into a PowerPoint presentation?” At that moment, you wish you
could press a button to automate the process because SAS programmers
spend way too much time updating PowerPoint slides. This paper offers
solutions to make your life easier by building upon techniques from Koen
Vyverman’s paper (SUGI 30, 2005) that discussed the Dynamic Data
Exchange (DDE) feature within SAS to write through MS Excel to MS
PowerPoint. The goal is to free data analysts from PowerPoint tyranny by
enabling efficient and repeatable PowerPoint presentations.
Business Intelligence Applications — Room 2009
9:00 a.m.
How Am I Driving - My Business? (Techniques, from the
Insurance Industry That Can Be Applied to Other
Business Areas to "Drive" Better Performance)
Guy Garrett, Achieve Intelligence
Steve Morton, Applied System Knowledge Ltd
(Invited) Paper 056-2013
This paper runs through the high level strategic measurements that general
insurance companies need to routinely monitor, looking at the technical
solutions available today using SAS® software. The techniques used in this
paper can also be applied to other industries helping executives and
managers to measure and monitor their businesses’ performance.
www.sasglobalforum.org/2013
69
Business Intelligence Applications — Room 3016
Business Intelligence Applications — Room 2009
9:00 a.m.
12:00 p.m.
Emerging Best Practices in the Age of Democratized
Analytics
Popular Tips and Tricks to Help You Use SAS® Web
Report Studio More Efficiently
Paper 541-2013
Paper 062-2013
Welcome to the age of analytics, everyone! Fact-based decision making has
never been more pervasive than today. Reports, dashboards and mobile BI
applications can empower entire organizations to understand and act on
analytics in the office and on the go. Your organization, team or division is
curious about data exploration capabilities and yet there are very diverse
levels of analytics understanding from person to person. Attend this
presentation for an overview of the best practices you can use to satiate the
curiosity of business decision makers to explore data, while maintaining
analytic integrity of your models and processes.
For more than six years, SAS® Web Report Studio has enabled users at all
skill levels to create, view, and explore centrally stored reports. This paper
discusses tips and tricks for reporting techniques that have been most
popular with customers over the years. The paper also explains new
features that shipped with the second maintenance release of SAS Web
Report Studio 4.31. As with the tips and tricks, these new features offer
more efficient methods for tasks related to conditional highlighting and to
content enhancement in reports that are sent via e-mail. The techniques
and features that are discussed cover tasks in the following key areas:
performance, filtering, scheduling, and distribution, report design, and
sending reports via e-mail.
Justin Choy, SAS
Keith Myers, SAS
Business Intelligence Applications — Room 2009
10:00 a.m.
How Mobile Changes the BI Experience
Murali Nori, SAS
Paper 053-2013
The advent of a new generation of tablets catapulted corporations’ use of
mobile devices. With SAS® Mobile BI for tablets, anyone who uses BI for
work and decision making has a new way to experience BI content. This
paper presents some end-to-end use cases to demonstrate how
revolutionary the user experience is with SAS Mobile BI. It also
demonstrates how easy it is to access and navigate BI content. Discover
how BI on mobile devices changes the user experience and the reach of BI
content for productivity, decision making and extracting better ROI.
Business Intelligence Applications — Room 3016
11:00 a.m.
What’s New in SAS® Enterprise Business Intelligence for
SAS® 9.3
Rick Styll, SAS
Paper 060-2013
SAS® Enterprise BI Server provides a comprehensive suite of BI tools that
allows a broad set of business and IT users to produce and consume
consistent, fact-based information. The latest revision contains
enhancements to both SAS Web Report Studio and SAS BI Dashboard. Key
capabilities are discussed and demonstrated by members of the product
team. Designing reports and dashboards is now more flexible, and
downstream consumers benefit from better performance, improved
navigation and interactions, and better integration with Excel and your
email client. Plans for future releases are previewed, such as mobile delivery
of SAS Web Report Studio reports, and how SAS Enterprise BI Server fits
within the overall BI portfolio.
70
www.sasglobalforum.org/2013
Customer Intelligence — Room 2001
9:00 a.m.
SAS® Treatments: One to One Marketing with
Customized Treatment Processes
Dave Gribbin, SAS
Amy Glassman, SAS
Paper 065-2013
The importance of sending the right message and offer to the right person
at the right time has never been more relevant than in today’s cluttered
marketing environment. SAS® Marketing Automation easily handles
segment-level messaging with out-of-the-box functionality. But how do
you send the right message and the appropriately valued offers to the right
person? And how can an organization efficiently manage many treatment
versions across distinct campaigns? This paper presents a case study of a
casino company that sends highly personalized communications and offers
to its clientele using SAS Marketing Automation. It describes how
treatments are applied at both a segment and one-to-one level. It outlines a
simple custom process that streamlines versioning treatments for reuse in
multiple campaigns.
10:00 a.m.
You’re Invited! Learn How SAS Uses SAS® Software to
Invite You to SAS® Global Forum
Lori Jordan, SAS
Shawn Skillman, SAS
Paper 066-2013
Discover how you are chosen to receive an invitation to SAS® Global Forum.
This paper explores the use of SAS® software, including SAS® Marketing
Automation, SAS® Enterprise Guide®, SAS® Enterprise BI, and SAS® DataFlux,
in the selection process. See how SAS Marketing uses strategic
segmentation practices and advanced analytics to target email
communications and improve list performance for SAS Global Forum.
11:00 a.m.
Financial Services — Room 2010
Hot off the Press: SAS® Marketing Automation 6.1
9:30 a.m.
Mark Brown, SAS
Brian Chick, SAS
Paper 067-2013
SAS® Marketing Automation 6.1 is a major release that introduces a new
browser-based user interface for business analysts. The new look and feel
deliver a highly interactive and intuitive user experience while delivering
key customer-driven features, including support for control groups, live
seeds, enhanced campaign definitions, and improved scheduling and
execution. This paper introduces new navigation, improved search
capabilities and easier sharing of information; it also presents easier reuse
of treatments, campaign components, scheduling and control group
methodologies. Various control group techniques will be presented,
including: - Hold-out control groups. - A/B testing control groups. Champion/challenger. - Challenger/challenger. Business case studies based
on customer examples illustrate the matching of campaign business goals
with the appropriate control group technique and why they are needed for
robust marketing measurement.
12:00 p.m.
Product Affinity Segmentation That Uses the Doughnut
Clustering Approach
Darius Baer, SAS
Goutam Chakraborty, Oklahoma State University
Paper 068-2013
Product affinity means the natural liking of customers for products. Product
affinity segmentation divides customers into groups based on purchased
products. While conceptually appealing to marketers and business analysts,
in practice it often yields inappropriate solutions such as one large segment
and many tiny segments. Standard transformations such as logarithms do
not help. In this paper, we demonstrate how a combination of softmax
transformations with a doughnut clustering approach (single central
cluster) results in more evenly sized product affinity segments for 30,000
customers of a business-to-business company. The affinity segments show
meaningful differences in product buying patterns across the customer
base, and can be used for identifying cross-selling and up-selling
opportunities. The segments are further profiled using customers'
background variables to provide deeper business insights.
12:30 p.m.
Predicting Women's Department Purchases in a Retail
Store By Using the SEMMA Methodology
Michael Soto, Ripley
Paper 069-2013
One of our focus areas is to improve the business in the Women's
Department because it is currently our most powerful department in terms
of transactions originated by customers. It raises the need to implement an
analytical model focused on this department for establishing what offer is
the most appropriate for our customers according to buying patterns of
customers, augmenting the likelihood that a customer comes back to the
stores. Those patterns are calculated based on demographic transactional
data, and any other interaction that our customers have had with our
stores. The predictive model we used is the logistic regression, and it was
executed following the SEMMA methodology considered by SAS® for
projects in SAS® Enterprise Miner™.
Managing and Analyzing Financial Risk on Big Data with
High-Performance Risk and Visual Analytics
Cary Orange, SAS
Donald Erdman, SAS
Stacey Christian, SAS
Paper 110-2013
SAS® High-Performance Risk is a distributed forecasting engine for financial
risk, used to compute such things as value at risk (VaR). The output created
by this engine can be voluminous and unwieldy since it represents future
portfolio prices, which can be billions of rows long. It is desirable to postprocess these results to produce ad hoc reports. For performance, it is
important to keep these results in parallel. This is the perfect situation to
use high-performance, in-memory solutions. In this paper, we present
examples and results of running SAS High-Performance Risk scenarios and
analyzing them with SAS Visual Analytics.
10:30 a.m.
A Case Study in Firmwide Stress Testing: Engineering
the CCAR Process
Carsten Heiliger, Sun Trust
(Invited) Paper 111-2013
Stress testing has become pervasive. The trouble lies in isolating the
substantive, insightful activity from the overwrought chaff. Almost an
overused platitude, the term “stress testing” can be inserted into just about
any process in a financial institution, and there will be an army of
consultants claiming to have a best-practice opinion on the topic. The
reality is far more convoluted. Often, what is referred to as a stress test is
simply a sensitivity analysis with a focus on a suboptimal outcome. Other
times, an operationally focused risk assessment is termed a stress test, as it
is analyzing processes performing below an optimal level.
11:30 a.m.
Integrated Framework for Stress Testing in SAS®
Jimmy Skoglund, SAS
Wei Chen, SAS
Paper 112-2013
Stress testing is an integrated part of enterprise risk management and is a
regulatory requirement. Stress testing is especially useful for integrating
forward-looking views into risk analysis. Indeed, stress tests can provide
useful information about a firm’s risk exposure that statistical risk methods,
calibrated on the basis of history, can miss. However, traditional stress
testing is done on a stand-alone basis. This makes the interpretation of risk
obtained from stress events vs. from risk analysis with statistical models
difficult to interpret. We consider a Markov model and innovative
implementation in SAS® that integrates rare stress events into regular
statistical risk models. The model allows a consistent integration of the
information in backward-looking historical data.
www.sasglobalforum.org/2013
71
12:30 p.m.
11:00 a.m.
Detecting Cross-Channel Fraud Using SAS®
Creating Clark Error Grid with SAS/GRAPH®, the SAS/
GRAPH Annotate Facility, and SAS® Macro Applications
Srikar Rayabaram, Oklahoma State University
Krutharth Kumar Peravalli Venkata Naga, Oklahoma State
University
Yongyin Wang, Medtronic Diabetes
John Shin, Medtronic Diabetes
Paper 113-2013
Paper 133-2013
In a world where criminals are getting effective in their ability to gain
information about a customer of a particular bank, cross-channel
monitoring and assessment has become very important. As each day
passes by, criminals are also getting bolder in terms of engaging beyond a
single channel to set in motion the movement of money. In these scenarios,
a cross-channel review of user activity is essential to detect or prevent
fraud. In this paper, we analyze data across various channels. Also, we
create a predictive model that can be used to predict such activity and
discuss how effective the model would have been to detect fraudulent
activity in the past.
Clarke Error Grid Analysis has been widely used in the accuracy
quantification of blood glucose values obtained from continuous glucose
monitoring (CGM) sensor against reference values from meter or YSI
instruments. A vivid graphic presentation of clinical accuracy of CGM sensor
data is preferred by statisticians and reviewers of regulatory agencies. SAS/
GRAPH® Annotate facility is a powerful tool for customizing, enhancing, or
changing the features of graphic outputs. Clarke Error Grid breaks down a
scatterplot of estimated glucose values versus reference values into five
zones: A, B, C, D, and E. This presentation demonstrates how to use SAS/
GRAPH, the SAS/GRAPH Annotate facility, and SAS macro applications
together to create such Error Grid for clinical accuracy determination of
CGM data against meter or YSI glucose values.
Foundations and Fundamentals — Room 2008
9:00 a.m.
Arrays - Data Step Efficiency
Harry Droogendyk, Stratia Consulting Inc.
Paper 519-2013
Arrays are a facility common to many programming languages, useful for
programming efficiency. SAS® data step arrays have a number of unique
characteristics that make them especially useful in enhancing your coding
productivity. This presentation will provide a useful tutorial on the rationale
for arrays and their definition and use.
9:30 a.m.
Creating ZIP Files with ODS
Jack Hamilton, Kaiser Foundation Hospitals
Paper 131-2013
ZIP files are a convenient way to bundle related files together, and can save
storage space at the same time. The ZIP format is used internally by SAS®
for SAS® Enterprise Guide® projects, but until SAS® 9.2 there was no native
way to create a ZIP file with your own SAS program. Starting in SAS 9.2, you
can create your own ZIP files using ODS PACKAGE statements. This
presentation describes how to create simple ZIP archives, and discusses
how to create an archive file with an internal directory structure.
10:00 a.m.
Three Easy Ways to Create Customized SAS® Graphs
Qinghua (Kathy) Chen, Gilead sciences Inc,
(Invited) Paper 132-2013
We often hear people saying that "a picture is worth a thousand words".
With that in mind, it basically tells us how powerful graphics can be when
used properly. Ways to make graphs with great visual impact has drawn a
great deal of attention from people in many fields and impactful graphics
help reviewers interpret the data. SAS® has made significant improvement
in graphs software over the past few years. With new features rolled out
such as Output Delivery System (ODS) graphics, Graphic Template
Language (GTL) and annotated data sets, creating customized graphics is as
easy as creating a simple plot. This paper will describe three easy ways to
create customized graphs in SAS.
72
www.sasglobalforum.org/2013
11:30 a.m.
Tips for Generating Percentages Using the SAS®
TABULATE Procedure
Kathryn McLawhorn, SAS
Paper 134-2013
PROC TABULATE is one of the few Base SAS® procedures that calculate
percentages. The procedure is unique in that it has many default statistics
for generating percentages and it provides the ability to customize
denominator definitions. Determining the right denominator definition is
an important, but often challenging, aspect of calculating percentages.
Written for intermediate users, this paper discusses techniques for
enhancing PROC TABULATE output with percentage statistics. Using
examples, the paper illustrates how to add standard percentages to the
PROC TABULATE output, and it shares tips for calculating percentages that
you might have thought not possible. The paper further illustrates how to
avoid common pitfalls that are related to structuring denominator
definitions and how to format table output.
12:30 p.m.
Developer Reveals: Extended Data Set Attributes
Diane Olson, SAS
Paper 135-2013
Have you ever wanted to save non-data information with your data set?
Now you can. Extended attributes allow you to store information related to
a data set or to a particular variable in a data set. Do you want to store the
SAS® code that created a particular data set? Do you need to save a URL
that specifies information about a particular variable in your data set? Do
you want to store a description of a variable or the formula used to produce
the variable value? This presentation by the developer shows you how to
do all of that and more with extended attributes.
Hands-on Workshops — Room 2011
Hands-on Workshops — Room 2011
9:00 a.m.
10:00 a.m.
SAS® Workshop: SAS® Data Integration Studio Basics
SAS® Workshop: SAS® Visual Analytics 6.1
Eric Rossland, SAS
Kari Richardson, SAS
Paper 534-2013
Paper 535-2013
This workshop provides hands-on experience using SAS Data Integration
Studio to construct tables for a data warehouse. Workshop participants will:
This workshop provides hands-on experience with SAS® Visual Analytics.
Workshop participants will:
• define and access source data
• explore data with SAS® Visual Analytics Explorer
• define and load target data
• design reports with SAS® Visual Analytics Designer
• work with basic data cleansing
11:00 a.m.
Hands-on Workshops — Room 2020
SAS® Workshop: SAS® Data Integration Studio Advanced
9:00 a.m.
Paper 536-2013
Create Your First SAS® Stored Process
Tricia Aanderud, And Data Inc
Angela Hall, SAS
Paper 148-2013
Learn how to convert a simple SAS® macro into three different stored
processes! Using examples from the newly released book “50 Keys to
Learning SAS Stored Processes,” you’ll see how to build a stored process
that allows users to filter their results for the report of their dreams. You’ll
learn how to use the SAS Prompt Framework to customize your stored
process quickly and efficiently. No experience required! Suitable for
beginners. SAS® 9.2 and later.
Kari Richardson, SAS
This workshop provides hands-on experience using a combination of
DataFlux Data Management Studio and SAS® Data Integration Studio.
Workshop participants will:
• Review two DataFlux Data Management Studio data jobs
• Upload the DataFlux Data Management Studio data jobs to the DataFlux
Data Management Server
• Review / create a SAS Data Integration Studio job that will execute the
uploaded data jobs on the DataFlux Data Management Server
Hands-on Workshops — Room 2020
11:00 a.m.
Hands-on Workshops — Room 2024
9:00 a.m.
So You're Still Not Using PROC REPORT. Why Not?
Ray Pass, PharmaNet/i3
Daphne Ewing, Auxilium Pharmaceuticals, Inc.
(Invited) Paper 149-2013
Everyone who can spell SAS® knows how to use PROC PRINT, and it
certainly has its place as a simple listing generator and as a debugging aid.
However, if a report generation/delivery tool with powerful formatting,
summarizing, and analysis features is called for, then PROC REPORT is the
solution. PROC REPORT can provide the standard PROC PRINT functionality,
but in addition, it can easily perform many of the tasks that you would
otherwise have to use the SORT, MEANS, FREQ, and TABULATE procedures
to accomplish. PROC REPORT is part of the Base SAS® product and can run
in either an interactive screen-painting mode or a batch mode. This handson workshop presents the basics of the batch (non-interactive) version of
PROC REPORT.
Ready to Become Really Productive Using PROC SQL?
Sunil Gupta, Gupta Programming
(Invited) Paper 150-2013
Using PROC SQL, can you identify at least four ways to select and create
variables, create macro variables, create or modify table structure, and
change table content? Learn how to apply multiple PROC SQL
programming options through task-based examples. This hands-on
workshop reviews topics in table access, retrieval, structure, and content, as
well as creating macro variables. References are provided for key PROC SQL
books, relevant webinars and podcasts, and key SAS technical papers.
Hands-on Workshops — Room 2024
11:00 a.m.
FREQ Out: Exploring Your Data the Old-School Way
Stephanie Thompson, Datamum
(Invited) Paper 151-2013
The tried-and-true FREQ procedure just doesn’t get the attention it
deserves. But, as they say, it is an oldie but a goodie. Sometimes you just
need a quick look at your data and a few simple statistics. PROC FREQ is a
great way to get an overview of your data with a limited amount of code.
This hands-on workshop explores everything from the basic framework of
the procedure to how to customize the output. It also presents an overview
of some of the options that are available.
www.sasglobalforum.org/2013
73
Hands-on Workshops — Room 2011
10:30 a.m.
12:00 p.m.
The Hospital Game: Optimizing Surgery Schedules to
Save Resources, and to Save Lives
SAS® Workshop: DataFlux® Data Management Studio
Basics
Kari Richardson, SAS
Paper 537-2013
This workshop provides hands-on experience using DataFlux® Data
Management Studio to profile then cleanse data. Workshop participants
will:
• learn to navigate DataFlux® Data Management Studio
• define and run a data profile
• define and run a data job
Operations Research — Room 2004
Andrew Pease, SAS
Ayesgul Peker, SAS
Paper 154-2013
Surgeons are required to perform vital operations on a daily basis, but often
their planning is not optimized for the “downstream” care. This leads to
under- or overutilization of postoperative nursing wards, which can either
compromise a hospital’s ability to provide the best possible care and ensure
the best possible patient outcome, or result in precious hospital funding
going toward underutilized resources. This paper briefly reviews some of
the academic research that is available for a data-driven, operations
research approach to solving this challenge, which is dubbed the “Hospital
Game” in some of this literature. The paper then proposes an optimizationbased approach that uses the OPTMODEL procedure to derive the best
surgery schedule for a major European hospital.
9:00 a.m.
11:00 a.m.
Using SAS® to Measure Airport Connectivity: An Analysis
of Airport Centrality in the US Network with SAS/IML®
Studio
Projecting Prison Populations with SAS® Simulation
Studio
Hector Rodriguez-Deniz, University of Las Palmas de Gran
Canaria
Pere Suau-Sanchez, Cranfield University
Augusto Voltes-Dorta, Universitat de Barcelona
(Invited) Paper 152-2013
The U.S. Federal Aviation Administration (FAA) estimates that $52.2 billion
will be available over the years 2011–2015 to fund airport infrastructure
developments. Because one of the main objectives is to reduce congestion
and delays, there is a need to acknowledge the importance of connectivity
(measured with a centrality indicator) when establishing funding priorities.
Currently, the FAA does not do this. In this paper, we exploit the capabilities
of SAS/IML® Studio to implement a range of centrality measures, construct
a graphical representation of the U.S. air transport network from airline
ticketing data, test the algorithms to identify hub airports, and study the
evolution of these indicators during the last decades in order to analyze the
impact of airline decisions on airport connectivity.
10:00 a.m.
Jeff Day, SAS
Ginny Hevener, NC Sentencing & Policy Advisory
Commission
Bahadir Aral, SAS
Tamara Flinchum, NC Sentencing & Policy Advisory
Commission
Emily Lada, SAS
Paper 155-2013
The majority of U.S. states are mandated to project prison populations for
the purpose of planning adequate capacity. Typical time series methods are
ineffective because they do not take into account factors like sentence
length, prior record, revocations, and legislative changes. Discrete event
simulation has proven to be a viable alternative. This paper discusses a
project in which SAS worked with the North Carolina Sentencing and Policy
Advisory Commission to build a model in SAS® Simulation Studio that
projects the number of prison beds needed for the next ten years. The
model uses current prison population data, recent court convictions,
revocations of community supervision, and estimates of growth to play out
the admissions and releases of inmates over the time horizon of the model.
Advanced Project Management beyond Microsoft
Project, Using PROC CPM, PROC GANTT, and Advanced
Graphics
Smarter Grid Operations with SAS/OR®
Paper 153-2013
Paper 156-2013
The Challenge: Instead of managing a single project, we had to craft a
solution that would manage hundreds of higher- and lower-priority
projects, taking place in different locations and different parts of a large
organization, all competing for common pools of resources. Our Solution:
Develop a Project Optimizer tool using the CPM procedure to schedule the
projects, and using the GANTT procedure to display the resulting schedule.
The Project Optimizer harnesses the power of the delay analysis feature of
PROC CPM and its coordination with PROC GANTT to resolve resource
conflicts, improve throughput, clearly illustrate results and improvements,
and more efficiently take advantage of available people and equipment.
Between the time electricity leaves utility generators and reaches your
home or business, 7% of the energy has been dissipated as heat. For the
average utility, this represents a total loss of more than $75 million each
year. Some of the electric current producing these losses does not result in
the actual production of power, and can be minimized with the proper
switching of devices that are located at strategic points in the distribution
system. These devices can also conserve energy by reducing voltages in the
distribution system and still provide a continuous supply of electricity to
the customer. This paper discusses the use of SAS/OR® to schedule device
switching to optimize the operations of the electrical distribution system.
Lindsey Puryear, SAS
Stephen Sloan, Accenture
74
www.sasglobalforum.org/2013
11:30 a.m.
Arnie de Castro, SAS
Greg Link, SAS
12:00 p.m.
9:30 a.m.
Vehicle Retail Forecasting Demand and Inventory
Management Case Study at Shanghai General Motors
Using the ADaM ADAE Structure for Non-AE Data
Paper 157-2013
(Invited) Paper 177-2013
This paper describes a case study about vehicle retail forecasting demand
and inventory management in the auto industry. It describes the project's
background and the problems that were addressed using SAS®.
The final and official ADaM ADAE structure titled “Analysis Data Model
(ADaM) Data Structure for Adverse Event Analysis” was developed as an
appendix to the ADaM v2.1 to allow simple production of standard Adverse
Event tables. An ADaM sub-team is expanding this structure to cover other
data analyzed in a similar fashion, such as Concomitant Medications. The
basic premise is that data with the same analysis needs as the standard
adverse events tables can and should use this structure. This presentation,
by members of that ADaM sub-team, describes the AE analysis need and
shows to apply it for other data, such as Concomitant Medications, Medical
History, and even Laboratory Events. Examples of ADaM SAS data sets, and
useful SAS® program code are included.
Christina Zhong, shanghai general motors
12:30 p.m.
Parallel Multistart Nonlinear Optimization with PROC
OPTMODEL
Ed Hughes, SAS
Tao Huang, SAS
Yan Xu, SAS
Paper 158-2013
Nonlinear optimization has many compelling applications, including
finance, manufacturing, pricing, telecommunications, health care,
engineering, and statistics. Often a nonlinear optimization problem has
many locally optimal solutions, making it much more difficult to identify a
globally optimal solution. That’s why the multistart feature in PROC
OPTMODEL selects a number of initial points and starts optimization from
each one, significantly improving your chances of finding a global
optimum. In SAS/OR® 12.1, the multistart feature adds parallel execution.
This paper explores the multistart feature and its parallel optimization
feature, illustrating with examples drawn from research and industry.
Pharma and Health Care — Room 2000
Sandra Minjoe, Octagon Research Solutions
Mario Widel, Roche Molecular Systems
10:30 a.m.
Developing Your SDTM Programming Toolkit
David Scocca, Rho, Inc.
Paper 178-2013
Data standards such as the Study Data Tabulation Model (SDTM) make
programmer’s lives simpler but more repetitive. The similarity across
studies of SDTM domain structures and relationships presents
opportunities for code standardization and re-use. This paper discusses the
development and use of tools to simplify the process of creating SDTM data
sets, with examples of common tasks and the code to implement those
tasks. It also discusses the usefulness of a metadata system and presents a
general specification for an interface for accessing metadata. Examples
include mapping study visits, parsing dates, and standardizing test codes.
9:00 a.m.
11:00 a.m.
Easy Button: A Process for Generating Standardized
Safety- and Non-Safety- Related Clinical Trial Reports
Assessing Drug Safety with Bayesian Hierarchical
Modeling Using PROC MCMC and JMP®
Xiangchen Cui, Vertex Pharmaceuticals, Inc.
Mominul Islam, Vertex Pharmaceuticals
Sanjiv Ramalingam, Vertex Pharmaceuticals Inc.
Jiannan Hu, Vertex Pharmaceuticals, Inc.
Yanwei Han, Vertex
Paper 176-2013
SAS has developed SAS® macros and template SAS programs based on its
standard (tables, figures, and listings) TFL shells for safety and non-safety
analysis. The new process includes developing reporting macros using
existing department macros to generate standard TFLs. The macros were
developed assuming the CDISC ADaM analysis data set standards, which
enable you to minimize the number of macro parameters for efficient use of
the macros by the user. The process shortens the development cycle time
and facilitates the adoption from SAS programmers to clinical reporting.
There is also a user manual and standard template programs. The process
reduces report generation time significantly and achieves the quality by
design principle.
Richard Zink, SAS
Paper 179-2013
Bayesian hierarchical models are advantageous for the analysis of adverse
events in clinical trials. First, the models can borrow strength across related
events within the MedDRA hierarchy. Second, the models can naturally
temper findings likely due to chance. We describe the implementation of
two Bayesian hierarchical models (Berry & Berry, 2004; Xia et al., 2010) used
for the analysis of adverse events using PROC MCMC. Once models are fit, it
is necessary to review convergence diagnostics to ensure that the posterior
samples of parameters sufficiently approximate the target distribution.
Numerous diagnostics are available within PROC MCMC, and we also
present a freely available JMP® add-in for MCMC (Markov Chain Monte
Carlo) dynamically interactive diagnostics, summary statistics and graphics.
12:00 p.m.
Predicting Health Care Expenditures with the MCMC
Procedure
Greg Watson, UCLA Center for Health Policy Research
Paper 496-2013
Substantial variation, excess zeros, skew and extreme outliers make fitting
and predicting health care expenditures rather difficult. This paper presents
a Bayesian model that uses the first year of the fourteenth panel
www.sasglobalforum.org/2013
75
(2009-2010) of the nationally representative Medical Expenditures Panel
Survey (MEPS) to predict health care expenditures for individuals in the
second year. The merits of a Bayesian approach are examined and
compared to classical alternatives. Implementation in the MCMC procedure
is presented in detail, and model diagnostics and validation are discussed.
12:30 p.m.
Doctoring Your Clinical Trial with Adaptive
Randomization: SAS® Macros to Perform Adaptive
Randomization
Jenna Colavincenzo, University of Pittsburgh
Paper 181-2013
Adaptive randomization schemes have become increasingly common in
beginning stages of clinical trials and in small clinical trials. This paper
introduces two kinds of adaptive randomization schemes (treatment
adaptive randomization and covariate adaptive randomization) and
discusses the benefits and limitations of each. In addition, this paper
demonstrates how to use SAS® macros to perform these adaptive
randomization schemes in a clinical setting, and how these macros can be
modified to fit your randomization needs.
Quick Tips — Room 2003
9:00 a.m.
A Macro to Verify a Macro Exists
Rick Langston, SAS
Paper 339-2013
Although the %SYSMACEXIST function can do macro existence checking, it
is limited to pre-compiled macros. This paper describes the
%MACRO_EXISTS macro, which verifies that a specified macro will be found
if invoked. The macro searches for pre-compiled macros as well as all
autocall libraries to verify their existence.
9:15 a.m.
A Simple Approach to Generate Page Numbers in X of Y
Format in ODS RTF Output
Amos Shu, Endo Pharmaceuticals
Paper 308-2013
Page numbers in X of Y format, such as "Page 18 of 280" is a common
feature of ODS RTF outputs. SAS borrows Microsoft Word processors to
compute those numbers and put them in the final output by using TITLE or
FOOTNOTE statements with "{page {\field{\fldinst{page}}} of
{\field{\fldinst{numpages}}}}" or "Page ~{thispage} of ~{lastpage}". However,
the page numbers generated by Microsoft Word processors contain field
code information displayed as "Page {PAGE \*MERGEFORMAT} of
{NUMPAGES \*MERGEFORMAT}" rather than the page numbers when Alt F9 keys are pressed. Some users such as medical writers do not like such
field code information. This paper discusses a simple way to generate page
numbers in X of Y format in ODS RTF output with the PROC REPORT
procedure.
76
www.sasglobalforum.org/2013
9:30 a.m.
A Macro to Read in Medi-Span Text Format Database by
Data Dictionary
Sijian Zhang, VA Pittsburgh Healthcar System
Paper 344-2013
Investigators often use commercial databases to obtain useful additional
information for their researches. However, many companies do not offer
the code for transferring the data files from their deliverable file format into
the one used in the customer’s system. With many data files and variables,
the data transfer process can be very tedious. If the databases vary in
different versions, the transfer code revision can be another pain. This
paper presents an approach to simplify the data transfer process of reading
in Medi-Span drug information text data files by taking the advantage of
macro programming and its data dictionary information. One of Medi-Span
text data files, “MF2STR”, is used as an example throughout this paper.
9:45 a.m.
%GetReviews: A SAS® Macro to Retrieve User Reviews in
JSON Format from Review Websites and Create SAS®
Data Sets
Siddhartha Reddy Mandati, Oklahoma State University
Ganesh Badisa, Oklahoma State University
Goutam Chakraborty, Oklahoma State University
Paper 342-2013
The proliferation of social networking sites and consumers’ desires to create
and share content on such sites has continued to generate a huge amount
of unstructured data. Analytics users often want to tap into such
unstructured data and extract information. Many websites such as Twitter,
Facebook, and Rotten Tomatoes offer APIs for external systems to interact
and retrieve the data in JSON format. The API of Rotten Tomatoes returns
data in a complex text pattern that has information about user reviews.
Currently, there is no designated code in SAS® to read the JSON response
directly and fetch the needed data. This paper illustrates the development
and application of a SAS Macro %GetReviews to retrieve the reviews of any
desired movie from Rotten Tomatoes’ API.
10:00 a.m.
Writing Macro Do Loops with Dates from Then to Now
Ronald Fehd, retired
Paper 343-2013
Dates are handled as numbers with formats in SAS® software. The SAS
macro language is a text-handling language. Macro %do statements
require integers for their start and stop values. This article examines the
issues of converting dates into integers for use in macro %do loops. Three
macros are provided: a template to modify for reports, a generic calling
macro function which contains a macro %do loop and a function which
returns a list of dates. Example programs are provided which illustrate unit
testing and calculations to produce reports for simple and complex date
intervals.
10:15 a.m.
11:00 a.m.
On a First-Name Basis with SAS: Creating Personalized
Error Messages Using SAS 9.2
SASY Codes for Lazy people
Andrew Clapson, Statistics Canada
Valerie Hastings, Statistics Canada
Paper 352-2013
In the interest of creating a user-friendly SAS® system, you might have the
good idea to include code that checks for common errors, notifies the user,
and suggests possible solutions. Apart from simply delivering this
information to the user, you might also use customized message windows
that express congratulations upon a successful run or even deliver lighthearted finger- wagging in the case of unexpected errors. Using SAS 9.2,
this paper details the steps necessary to include basic error messaging
functionality in SAS programs. It covers notification of specific errors as well
as confirmation of successful program execution. In addition, through the
use of system macro variables, these feedback messages can surprise users
by ‘knowing’ their names and addressing them directly.
10:30 a.m.
Graph Your SAS® Off
Karena Kong, InterMune
Paper 309-2013
This paper demonstrates three different SAS® procedures for creating
graphs. For illustration purposes, the bubble plot in Figure 1, shows the
ratio of broadband users (DSL, Cable, Other) ranked by population ("List of
countries,").The data values of “Total Subscribers in Millions” and “Percent
Population Online” are annotated on the graph. The three procedures are
from SAS/GRAPH® - GPLOT, Statistical Graphics (SG) â SGPLOT and Graphics
Template Language (GTL) - PROC TEMPLATE with SGRENDER. This paper
will discuss the advantages and disadvantages between each one. Based on
the comparisons, it recommends which procedure should be used to create
a similar graph.
10:45 a.m.
Using SAS® to Assess Individuals’ Best Performance in
Multiple Dimensions
Aude Pujula, Louisiana State University
David Maradiaga, Louisiana State University
Paper 346-2013
There are many cases where we need to look at the best performance of an
individual in several disciplines over multiple time events. For instance, we
might want to know a triathlete’s best position in the three disciplines over
all the races of the season, or the highest test scores of a student in several
sub-scores. Looking at the latter example, this paper compares four
different methods implementable in Base SAS® to create a data set that
contains one record per student corresponding to the highest test scores.
Of particular interest is the use of PROC SQL combined with the SELECT
DISTINCT clause and the MAX function that allows the creation of the
desired data set in one step.
Prashanthi Selvakumar, UNT Health Science Center
Paper 353-2013
"I choose a lazy person to do a hard job, because a lazy person will find an
easy way to do it." - Bill Gates. Everyone wants to save time. While hard
work is useful, smart work is a pre requisite. Are you tired of typing codes,
then read this paper, it gives you the ways to shorten your codes. The topics
discussed in this paper include, array, do loops, macros, functions. It also
discusses the procedures and data steps where macros can save your time.
The other techniques like, combining the macros while creating html, pdf,
rtf output, to produce professional report. The possible ways of saving time
in programming are addressed in this paper.
11:15 a.m.
The SAS® Versus R Debate in Industry and Academia
Chelsea Lofland, University of California, Santa Cruz
Rebecca Ottesen, City of Hope and Cal Poly State University,
San Luis Obispo
Paper 348-2013
Despite industry being heavily dominated by SAS®, R is used widely in
academia due to being free and open-source software that is structured
around users being able to write and share their own functions. However,
this disconnect leaves many students who are pursuing analytic degrees
struggling to get a job with less SAS experience than desired by companies.
Alternatively, they could face the struggle of transitioning everything they
learned in university from R to SAS. Ideally, one would know every possible
programming language and use the one that best suits the situation. This is
rather unrealistic. Our goal is to show the benefits of these two very
different software packages and how to leverage both of their strengths
together.
11:30 a.m.
Using SAS® to Dynamically Generate SAS® Code in Order
to Display Both Variable Label and Name as Column
Header in PROC REPORT and PROC PRINT
Victor Lopez, Baxter Healthcare Corporation
Heli Ghandehari, Baxter BioScience
Paper 349-2013
With implementation of data standards such as CDISC SDTM, datasets
contain sufficiently meaningful variable names and labels to allow direct
reporting from dataset to output (PDF, RTF, and many more). This
eliminates the necessity to program lengthy DEFINE statements in PROC
REPORT or to manually assign custom labels in PROC PRINT. This paper
illustrates an innovative approach using SAS® to dynamically generate SAS
code that enables us to solve a seemingly easy problem: displaying both
the variable label and name as a column header in PROC REPORT and PROC
PRINT.
www.sasglobalforum.org/2013
77
11:45 a.m.
9:30 a.m.
Best Practices: PUT More Errors and Warnings in My Log,
Please!
Extending SAS® Reports to Your iPhone
Koketso Moeng, Statistics South Africa
Mary Rosenbloom, Edwards Lifesciences, LLC
Kirk Paul Lafler, Software Intelligence Corporation
Paper 350-2013
We all like to see a SAS® log that is free from errors and warnings, but did
you know that you can add your own errors and warnings to the log with
PUT statements? Not only that, but you can incorporate this technique into
your regular coding practice to check for unexpected data values. This
paper will explore the rationale and process of issuing user-created error
and warning messages to the SAS log, along with a number of examples to
demonstrate when this is useful. Finally, we will propose an upgrade to the
next version of SAS involving a user-specified keyword with its own color in
the log.
Paper 378-2013
You have jumped through all of the hoops of creating the perfect
dashboards for executives, marketing, human resources, finance, and the
project office teams, but they hardly ever get used because, frankly, your
users don't have enough time in the day to go through the reports. This is
even more true if they have to be tethered to the servers in the office to do
so. Luckily, a solution that suits the user with an iPad or iPhone is available.
Introducing Roambi—a mobile business intelligence (BI) platform that runs
on Apple's iOS platform and can be easily integrated into your existing SAS®
Enterprise BI platform.
10:00 a.m.
12:00 p.m.
The Dynamic Cube Viewer - OLAP Made Easy
The Surprisingly "Sym"ple Alternative to Hardcoding
Paper 379-2013
Rachel Carlson, Mayo Clinic
Ruchi Sharma, Mayo Clinic
Paper 351-2013
As a frequent SAS® user, do you often feel that you are spending too much
time looking up procedure results or hardcoding the values into programs?
Does your data often change causing a need to rerun analyses, forcing you
to repeat steps? Save time rerunning analysis programs by reducing the
amount of hardcoded variables and formats. Our paper will demonstrate
how to effectively use the CALL SYMPUTX routine and the SYMGET function
to make your code more flexible and minimize the possibility of data
calculation errors.
Reporting and Information Visualization — Room
2002
9:00 a.m.
Horizontal Data Sorting and Insightful Reporting: A
Useful SAS® Technique
Justin Jia, CIBC
Amanda Lin, Bell Canada
Paper 376-2013
Sorting and ordering of data is a fundamental skill in SAS® data analysis.
Data sorting can be vertical sorting, across rows, or horizontal sorting,
across columns. Compared to vertical sort, horizontal sort is used less
frequently, and it requires the user to employ multiple sophisticated SAS
skills such as Transpose, Rotate, Array, Macro, etc. It is also an important and
useful technique for advanced data analysis and reporting in customer
profiling and metrics, which can significantly enhance the format and
layout of data reporting, and thus provide informative insights into data.
This paper will discuss the different approaches and methods of performing
horizontal sorting and presentation of SAS data, which can also expand our
horizon on data manipulation and SAS programming skills.
78
www.sasglobalforum.org/2013
Raymond Ebben, OCS Consulting
The Dynamic Cube Viewer is a bespoke browser-based application that
offers an intuitive interface for business users to query OLAP cubes, without
the need to have an understanding of OLAP cubes. It has originally been
developed as a benchmarking tool for the Association of Dutch Insurers
and has been further developed by OCS Consulting to make it more
generic. The application reads only OLAP cube metadata and uses this to
build the user interface. An impression can be found in the attached
abstract.
10:30 a.m.
Statistical Graphics for Clinical Research Using ODS
Graphics Designer
Wei Cheng, Isis Pharmaceuticals, Inc.
Paper 380-2013
Statistical graphics play an important role across various stages in clinical
research. In this paper, I will show you the application interface and walk
you through creating some commonly used statistical graphs for clinical
research. The intended audience doesn’t need to know SAS/GRAPH®
syntax, but wants to create high-quality statistical graphs for clinical trials.
Examples will use scrambled data from real world in CDISC format.
11:00 a.m.
Visualize Your OLAP Cubes on a Map through a Stored
Process
Frank Poppe, PW Consulting
Sjoerd Boogaard, Kiwa Prismant
Paper 381-2013
What if you have an OLAP cube with a geo-dimension and you want a map
from that, but you don't have ArcGIS? Enter this general stored process. It
can read measures and dimensions from the cube, and it uses SAS/GRAPH®
software to combine that with boundary data, creating a color-coded map.
The map is clickable to navigate between the geographical levels. A
selection pane offers measures, and non-geographical dimensions surface
as (hierarchical) filters. Measures and dimensions are read from the
metadata; values for the filters are read from the cube, using MDX.
Boundaries are clipped to the right zoom level, and the picture gets a
background with roads from a web service. HTML, CSS, and JS is generated
to glue everything together and to deliver it to the portal.
11:30 a.m.
Retail — Room 3014
GTL to the Rescue!
8:00 a.m.
Paper 382-2013
Location Planning: A Look into Location Planning - Best
Practices for Resolving
Lelia McConnell, SAS
You just produced some graphs using the SGPLOT and SGPANEL
procedures. Now you want to modify the structure of your graphs in order
to make them more meaningful. You are looking for options that enable
you to split the axis values across multiple lines, add a table under a graph,
or create a template where you can conditionally execute statements and
dynamically assign variables. However, you cannot find any options within
the procedure that enable you to put these final touches on your graphs.
When all seems hopeless, ODS Graphics Template Language (GTL) comes to
the rescue!
12:00 p.m.
SCAD: Development of Statistical Information Systems
for the Provision of Census Data
Greg Pole, Statistics Centre Abu Dhabi
Paper 356-2013
The Statistics Centre - Abu Dhabi (SCAD) was founded in 2008 and seeks to
join the world’s leading statistical organizations in statistical collection,
production, and dissemination. In October 2011, SCAD conducted its first
census of population and households. In addition to using innovative
enumeration technologies (e.g., iPads), SCAD is also advancing the
development of inventive and flexible tools for accessing rich census data.
This is a positive shift towards greater access to public data in the Emirate.
The tools SCAD has developed for the 2011 Census use SAS® as a
foundation and include: on-line Thematic Mapping, on-line Community
Tables, and on-line Table Builder. These tools will be released to the Abu
Dhabi government and public in 2012, as web-based applications.
Reporting and Information Visualization — Room
3016
12:00 p.m.
Ann Ferguson, SAS Institute
(Invited) Paper 392-2013
Location planning is a struggle to balance the trends of the merchandise
and stores while meeting the financial objectives of the company. The Store
Planner, plans the sales forecast and sales growth for the each and every
category available in the store or location. The number of stores in most
retail organizations is typically large and developing the store level plans is
a voluminous task. This panel discussion features recent retail trends and
efforts to maximize profits and drive improvements. Hear how these
retailers are using SAS and innovative methods and tools to develop plans
tailored for merchandise and location trends.
9:00 a.m.
Implementing Assortment Planning and the challenge of
User Adoption
Ann Ferguson, SAS Institute
(Invited) Paper 393-2013
Merchandise Assortment planning plays a pivotal role in creating and
maintaining profitability. No other area within a retail business has such a
direct impact on bottom line profit (or loss). It is, therefore, crucial that
merchandisers have a broad understanding of the best practice approaches
that have evolved and continue to evolve in order that they are able to
optimize the financial return on the investment that is under their control.
10:00 a.m.
Reclass: What Does It Mean to You!
Amy Clouse, Dick's Sporting Goods
(Invited) Paper 394-2013
“Google-Like” Maps in SAS®
Reclass: What does it mean to you! Reclass can be a daunting task to take
on in your organization. This session will focus on best practices for
preparation and execution, compiled from several SAS® customers.
Paper 377-2013
11:00 a.m.
Darrell Massengill, SAS
We are frequently asked if we can have maps similar to Google Maps in
SAS®. Customers want the background image displayed behind their data
so they can see where streets or other features are located. They may also
want to pan and zoom the map. Unfortunately, Google has legal
restrictions and limitations on the use of their maps. Now, you can have
“Google-like” maps inside of SAS. You may have already seen this capability
in products like SAS® Visual Analytics Explorer and other products using
them will be available in future releases. This presentation will discuss and
demonstrate these new capabilities in SAS Visual Analytics Explorer, SAS/
GRAPH®, and other products.
Forecasting to Support Planning
Julie Rankin, Belk
(Invited) Paper 395-2013
Strength in Numbers: Using Demand Forecasting to Drive Merchandise and
Store Performance Hear how these retailers are using SAS Demand
Forecasting for Retail to strengthen their numbers through advanced retail
management from the SAS forecast engine in their planning processes.
With broad assortment and locations, diverse consumer behavior and price
fluctuations, predicting demand and performance can be quite a challenge
in the pre-season and in-season planning processes. These retailers are
using analytics to drive results and supplement the art of the merchant
expertise to generate efficiencies via a scientific approach to forecasting
demand. Forecast results drive better business decisions, improved
planning processes and forecast accuracy. The user experiences reflect
streamlined processes, improved productivity and better demand patterns.
www.sasglobalforum.org/2013
79
12:00 p.m.
SAS® Retail Planning 7.2 Demonstration
Elaine Markey, SAS
Paper 396-2013
A live demonstration of SAS Retail Planning 7.2. This demonstration will
show how this release helps retailers plan their financials and organize
customer-centric localized assortments in an effective and efficient manner.
SAS® Enterprise Guide® Implementation and
Usage — Room 3002
9:00 a.m.
SAS® Enterprise Guide®, Best of Both Worlds: Is it Right
for You?
Sunil Gupta, Gupta Programming
(Invited) Paper 416-2013
SAS and Big Data — Room 3001
9:00 a.m.
High Performance Statistical Modeling
Bob Rodriguez, SAS
Robert Cohen, SAS
Paper 401-2013
The explosive growth of data, coupled with the emergence of powerful
distributed computing platforms, is driving the need for high-performance
statistical modeling software. SAS has developed a series of procedures that
perform statistical modeling and model selection by exploiting all of the
cores available – whether in a single machine or in a distributed computing
environment. This presentation includes a demonstration of the current
capabilities of this software as well as guidance on how and when these
high-performance procedures will provide performance benefits.
10:00 a.m.
SAS® Visual Analytics Road Map
Greg Hodges, SAS
Paper 513-2013
Journey down the SAS® Visual Analytics road map with product
management! Learn about the reporting, exploration and even new data
visualizations under consideration for future releases.
12:00 p.m.
Getting Started with SAS® Visual Analytics
Administration and Deployment
Meera Venkataramani, SAS
Gary Mehler, SAS
Paper 515-2013
Discover invaluable tips on how to get started with SAS® Visual Analytics
deployment, administration and monitoring. Hear tips and best practices
for a flawless implementation.
80
www.sasglobalforum.org/2013
Whether you are new to SAS® or a seasoned SAS Programmer, you still face
the same dilemma. Does SAS® Enterprise Guide® represent the best of both
worlds to make the transition to SAS easier with a point-n-click interface or
enhance your productivity with over 90 tasks? Do you follow the same
traditional path taken by millions who learned SAS many decades ago or do
you take the yellow brick road to directly analyze your data? This
presentation explores the vast differences between these two cultures and
how they impact your programming environment. While there are
numerous benefits to using SAS Enterprise Guide, there are also some
caveats to keep in mind to make the transition smoother.
10:00 a.m.
Update, Insert, and Carry-Forward Operations in
Database Tables Using SAS® Enterprise Guide®
Thomas Billings, Union Bank
Paper 417-2013
You want to use SAS® Enterprise Guide® to simulate database logic that
includes any of: update, insert, carry-forward operations on old, changed, or
new rows between two data sets, to create a new master data set. However,
the Query Builder Task GUI does not have an Update/Insert option.
Methods for simple types of update, insert, and/or carry-forward operations
are described and illustrated using small data sets. First, we review Base
SAS® methods, including DATA step and PROC SQL code. Then, two GUIonly/Task-based methods are described: one based on the Sort Data Task
GUI; the other on the Query Builder Task GUI. The issue of whether integrity
constraints are preserved is also discussed.
10:00 a.m.
Update, Insert, and Carry-Forward Operations in
Database Tables Using SAS® Enterprise Guide®
Sreenivas Mullagiri, iGATE Global Solution
Paper 417-2013
You want to use SAS® Enterprise Guide® to simulate database logic that
includes any of: update, insert, carry-forward operations on old, changed, or
new rows between two data sets, to create a new master data set. However,
the Query Builder Task GUI does not have an Update/Insert option.
Methods for simple types of update, insert, and/or carry-forward operations
are described and illustrated using small data sets. First, we review Base
SAS® methods, including DATA step and PROC SQL code. Then, two GUIonly/Task-based methods are described: one based on the Sort Data Task
GUI; the other on the Query Builder Task GUI. The issue of whether integrity
constraints are preserved is also discussed.
10:30 a.m.
Statistics and Data Analysis — Room 2005
A Comparison between GUI Prompts of SAS® Enterprise
Guide® 4.1 and 4.3 and Approaches for Developing NextGeneration Prompts
9:00 a.m.
Menaga Ponnupandy, Technosoft Corp
Paper 418-2013
SAS® codes have to be edited when the criteria of execution changes. The
use of GUI Prompts helps in preserving source code from changes and in
automation of SAS® Enterprise Guide® Projects by passing run-time
parameters to SAS. The main purpose of this paper is to compare the
advanced features or functionalities of GUI Prompts between SAS
Enterprise Guide 4.1 and SAS Enterprise Guide 4.3. This paper also discusses
about the limited ability of prompts and provides tips for handling such
situations or highlights the need for development of future generation
prompts.
11:00 a.m.
Making do with less: Emulating Dev/Test/Prod and
Creating User Playpens in SAS® Data Integration Studio
and SAS® Enterprise Guide®
David Kratz, d-Wise
Paper 419-2013
Have you ever required a Dev / Test / Prod environment but found yourself,
for whatever reason, unable to lay down another SAS Installation? Have you
ever discovered that your results have been overwritten by a team
member? Our ability to use SAS is shaped by the environment in which the
software is installed, but we often don't have as much control over that
environment as we'd like. However, we can often emulate the setup we'd
prefer by configuring the one we have. This paper explores this concept
using techniques which can be applied to development in SAS Data
Integration Studio and SAS Enterprise Guide.
11:30 a.m.
What SAS® Administrators Should Know About Security
and SAS® Enterprise Guide®
Casey Smith, SAS
Paper 420-2013
SAS® Enterprise Guide® is a flexible and powerful tool in the hands of your
users. However, as every superhero knows, with power comes
responsibility. As an administrator, you want to ensure various groups of
users have access to the specific resources they need to be as productive as
possible, while at the same time protecting company assets and minimizing
risk. This paper explores various security considerations from the SAS
Enterprise Guide perspective, such as authentication, authorization, user
administration, access management, encryption and role-based availability
of application features.
Estimating Censored Price Elasticities Using SAS/ETS®:
Frequentist and Bayesian Approaches
Christian Macaro, SAS
Jan Chvosta, SAS
Kenneth Sanford, SAS
James Lemieux, SAS Institute
Paper 448-2013
The number of rooms rented by a hotel, spending by “loyalty card”
customers, automobile purchases by households—these are just a few
examples of variables that can best be described as “limited” variables.
When limited (censored or truncated) variables are chosen as dependent
variables, certain necessary assumptions of linear regression are violated.
This paper discusses the use of SAS/ETS® tools to analyze data in which the
dependent variable is limited. It presents several examples that use the
classical approach and the Bayesian approach that was recently added to
the QLIM procedure, emphasizing the advantages and disadvantages that
each approach provides.
Statistics and Data Analysis — Room 2007
9:00 a.m.
Considerations and Techniques for Analyzing Domains
of Complex Survey Data
Taylor Lewis, U.S. Office of Personnel Management
(Invited) Paper 449-2013
Despite sounding like a straightforward task, making inferences on a
domain, or subset, of a complex survey data set is something that is often
done incorrectly. After briefly discussing the features constituting complex
survey data, this paper explains the risks behind simply filtering the full data
set for cases in the domain of interest prior to running a SAS/STAT® survey
procedure such as PROC SURVEYMEANS or PROC SURVEYREG. Instead, it
shows how one should use the DOMAIN statement or create a domainspecific analysis weight. Also discussed in detail are considerations and
approaches to the very common objective of testing whether the
difference between two domain means is statistically significant.
Statistics and Data Analysis — Room 2005
10:00 a.m.
Markov Chains and Zeros in My Data: Bayesian
Approaches in SAS® That Address Zero Inflation
Matthew Russell, University of Minnesota
Paper 450-2013
In recent releases of SAS/STAT® software, a number of procedures that
perform Bayesian methodologies have been incorporated. A common
modeling problem across many disciplines is that of addressing largerthan-expected proportions of zeros, a problem that is exacerbated when
counts and probabilities of zeros are heterogeneous. This paper uses
examples from the ecological literature to perform Bayesian analyses on
discrete data with zero inflation. We focus primarily on the MCMC
procedure, but also address use of Bayesian methods in the FMM and
www.sasglobalforum.org/2013
81
GENMOD procedures. We fit zero-inflated models under conditional
binomial, Poisson, and negative binomial assumptions both with and
without random intercept effects.
illustrates the basic framework of an LGC model and introduces a SAS
macro, %LGCM, that fits a latent growth model and computes incremental
fit indices based on more appropriate baseline models.
Statistics and Data Analysis — Room 2007
Statistics and Data Analysis — Room 2005
10:00 a.m.
11:00 a.m.
Exploring Health Trends and Risk Behavior Analysis in
American Youth Using PROC SURVEYFREQ and PROC
SURVEYLOGISTIC
The Box-Jenkins Methodology for Time Series Models
Deanna Schreiber-Gregory, North Dakota State University
Paper 451-2013
This study looks at recent health trends and behavior analyses of youth in
America. Data used in this analysis was provided by the Centers for Disease
Control and Prevention and gathered using the Youth Risk Behavior
Surveillance System (YRBSS). This study outlines demographic differences
in risk behaviors, health issues, and reported mental states. Interactions
between risk behaviors and reported mental states were also analyzed.
Visual representations of frequency data for the national results are also
provided and discussed. A final regression model including the most
significant contributing factors to suicidal ideation is provided and
discussed. Results included reporting differences between the years 1991
and 2011. All results are discussed in relation to current youth health trend
issues. Data was analyzed using SAS® 9.3.
Theresa Ngo, Warner Bros. Home Entertainment
Paper 454-2013
A time series is a set of values of a particular variable that occur over a
period of time in a certain pattern. The most common patterns are
increasing or decreasing trend, cycle, seasonality, and irregular fluctuations
(Bowerman, O’Connell, and Koehler 2005). To model a time series event as a
function of its past values, analysts identify the pattern with the assumption
that the pattern will persist in the future. Applying the Box-Jenkins
methodology, this paper emphasizes how to identify an appropriate time
series model by matching behaviors of the sample autocorrelation function
(ACF) and partial autocorrelation function (PACF) to the theoretical
autocorrelation functions. In addition to model identification, the paper
examines the significance of the parameter estimates, checks the
diagnostics, and validates the forecasts.
Statistics and Data Analysis — Room 2007
Statistics and Data Analysis — Room 2005
11:00 a.m.
10:30 a.m.
Introducing the New ADAPTIVEREG Procedure for
Adaptive Regression
Forecasting Net Job Creation Using SAS®
Casey Sperrazza, University of Alabama
Paper 453-2013
Using data from the U.S. Census Bureau’s Business Dynamics Statistics, net
job creation is forecast economywide and by sector. Forecasts are carried
out economywide using exponential smoothing and ARIMA models.
Forecasting is carried out by Census Bureau–defined sectors using ARIMA
models. Data are from 1977–2010, and net job creation is forecast through
2020.
Statistics and Data Analysis — Room 2007
10:30 a.m.
Modeling Change over Time: A SAS® Macro for Latent
Growth Curve Modeling
Pei-Chin Lu, University of Northern Colorado
Robert Pearson, University of Northern Colorado
Paper 452-2013
In recent years, latent growth curve (LGC) modeling has become one of the
most promising statistical techniques for modeling longitudinal data. The
CALIS procedure in SAS® 9.3 could be used to fit an LGC model. As one
application of structural equation modeling (SEM), LGC modeling relies on
indices to evaluate model fit. However, it has been pointed out that when
you are obtaining incremental fit indices, the default baseline model used
in many popular SEM software packages, including PROC CALIS, is generally
not appropriate for LGC models (Widaman and Thompson 2003). This paper
82
www.sasglobalforum.org/2013
Warren Kuhfeld, SAS
Weijie Cai, SAS
Paper 457-2013
Predicting the future is one of the most basic human desires. In previous
centuries, prediction methods included studying the stars, reading tea
leaves, and even examining the entrails of animals. Statistical methodology
brought more scientific techniques such as linear and generalized linear
models, logistic regression, and so on. In this paper, you will learn about
multivariate adaptive regression splines (Friedman 1991), a nonparametric
technique that combines regression splines and model selection methods.
It extends linear models to analyze nonlinear dependencies and produce
parsimonious models that do not overfit the data and thus have good
predictive power. This paper shows you how to use PROC ADAPTIVEREG (a
new SAS/STAT® procedure for multivariate adaptive regression spline
models) by presenting a series of examples.
Statistics and Data Analysis — Room 2005
11:30 a.m.
Exploring Time Series Data Properties in SAS®
David Maradiaga, Louisiana State University
Aude Pujula, Louisiana State University
Hector Zapata, Louisiana State University
Paper 456-2013
Box and Jenkins popularized graphical methods for studying time series
properties of time series data. Dickey and Fuller did the same for unit root
tests. Both methods seek to understand the nonstationary properties of
improvement in overall throughput and has allowed eBay Inc in ~30%
additional processing capacity, and thereby enabled evaluating more
experiments.
data, and SAS® software is a popular tool used by applied researchers. The
purpose of this paper is to provide a series of steps using the SAS macro
language, PROC SGPLOT, PROC ARIMA, PROC AUTOREG, and the %dftest
macro to diagnose nonstationary properties of data. A comparison of three
competing SAS procedures is presented, with SAS capabilities highlighted
using simulated time series.
10:30 a.m.
12:00 p.m.
Hardening a SAS® Installation on a Multi Tier installation
on Linux
Nontemporal ARIMA Models in Statistical Research
David Corliss, Magnify Analytic Solutions
(Invited) Paper 458-2013
Mathematical models employing an autoregressive integrated moving
average (ARIMA) have found very wide applications following work by Box
and Jenkins in 1970, especially in time series analysis. ARIMA models have
been very successful in financial forecasting, forming the basis of such
things as predicting how much gas prices will rise. However, no
mathematical requirement exists requiring the data to be a time series: only
the use of equally spaced intervals for the independent variable is
necessary. This can be done by binning data into standard ranges, such as
income by $10,000 intervals. This paper reviews the fundamental statistical
concepts of ARIMA models and applications of non-temporal ARIMA
models in statistical research. Examples and applications are given in
biostatistics, meteorology, and econometrics as well as astrostatistics.
Jan Bigalke, Allianz Managed Operations & Services SE
Paper 481-2013
The security requirements of today require in some use cases the hardening
of a SAS® Installation. This paper describes the practical steps of securing
the SAS web applications and the impact to the Base SAS® Services on the
SAS computer tiers. The SAS® Enterprise BI Server will be the object of this
explanation. The principles of a secure architecture will be described and
the options to secure the individual components presented.
11:00 a.m.
Do I Need a Migration Guide or an Upgrade Coach?
Donna Bennett, SAS
Mark Schneider, SAS
Gerry Nelson, SAS
Paper 482-2013
Systems Architecture and Administration — Room
2006
9:00 a.m.
Benchmarking SAS® I/O: Verifying I/O Performance
Using fio
Spencer Hayes, J. S. Hayes, Inc.
Paper 479-2013
Input/Output (I/O) throughput is typically the most important computing
aspect of a SAS® environment. Bandwidth requirements ranging from
25MB/sec/core to 135MB/sec/core are common in a high-performance SAS
system. Insuring that the storage subsystem can meet the demands of SAS
is critical to delivering the performance required by the business and user
community. Ideally, SAS administrators could run real-world SAS jobs to
benchmark the I/O subsystem. However, technical and logistical challenges
frequently make that option impractical. The open source software tool
called “fio” provides a method for accurately simulating I/O workloads. It is
configurable to match closely the existing or expected I/O profile for a SAS
environment.
9:30 a.m.
eBay Quadruples Processing Speed with SAS® InDatabase Analytics for Teradata
John Scheibmeir, eBay
(Invited) Paper 480-2013
Don’t bring a hammer when you need a paintbrush! There are many kinds
of changes you can make to your SAS® deployment. Sometimes, SAS
migration tools may provide the best path for making changes to your
system. Other times, your changes may need other deployment and
management tools. Whether you are an administrator managing the
changes or an IT administrator overseeing SAS software, this paper will help
you choose the right tools to plan and manage SAS software changes. As an
added bonus, the paper includes a glossary of common terms and concepts
that often require collaboration between IT and SAS management.
11:30 a.m.
Integrating SAS® into Your Operational Environment:
SOA: A Means to an End
Saravana Chandran, SAS
Rob Stephens, SAS
Paper 483-2013
The impact of business analytic models has proven value for enterprises.
Many SAS customers have highly valuable analytical assets, ranging from
analytical models to analytical services specific to domain. We have seen a
significant stream of requests for assistance in taking the next step to
deploy these models into their primary operational business applications.
Service-oriented architecture (SOA) is an architectural style designed to
enable flexibility, reusability and interoperability; it provides one of the
primary means for integrating SAS® with your operational application
environment. The paper walks through the integration, runtime
environment, governance and best practices all in the context of SOA and
SAS Business Analytics.
Working efficiently with HUGE data sets consisting of millions of rows and
hundreds of columns summing up to gigabytes of storage is a challenge
that many users and organizations face today. In addition to processing
large amounts of data, additional constraints include end-to-end
processing time, implications of transfer of processing to the database,
storage space, system resources, data transfer, etc. Utilizing SAS® indatabase processing on eBay’s Teradata based Singularity Platform has
reduced end-to-end processing time by a factor of 4 at eBay Inc. This
www.sasglobalforum.org/2013
83
12:00 p.m.
How to Choose the Best Shared File System For Your
Distributed SAS® Deployment
Barbara Walters, SAS
Ken Gahagan, SAS
Leigh Ihnen, SAS
Vicki Jones, SAS
Paper 484-2013
A shared file system is an integral component of all SAS® Grid Manager
deployments, SAS Enterprise BI deployments with load balanced servers on
multiple systems, and other types of distributed SAS applications. This
paper explains how SAS software interacts with the file system and how a
shared file system behaves differently than a non-shared file system. It
describes the factors that determine whether a particular file system is a
good fit for your environment and how to choose the file system that best
meets your needs.
12:30 p.m.
Bridging the Gap Between SAS® Applications Developed
by Business Units and Conventional IT Production
Thomas Billings, Union Bank
Euwell Bankston, Union Bank, NA
Paper 485-2013
Multiple factors are involved in the decision by an enterprise to decide
whether to allow a business unit to run its own production versus having
SAS® applications developed by business units run in conventional IT
production. There can be a wide gap between the business unit view of
"production-ready" programs vs. core IT standards for production systems.
The nature of the gap is discussed here, and also the risks of business-run
production. Specific suggestions are made regarding whether IT and
business should have joint ownership of critical SAS applications vs.
segregated roles, and when/how should SAS-based systems be migrated
into a fully controlled IT production environment.
84
www.sasglobalforum.org/2013
A
Aanderud, Tricia 8 , 16 , 61 , 73
Abousalh-Neto, Nascif 41
Achanta, Bhargav 21
Agrawal, Gaurav 41
Akerman, Meredith 26
Albright, Russell 44 , 59
Alexander, Malcolm 8 , 42
Amezquita, Darwin 43
Anderson, Brett 16
Anderson, Marty 58
Ansari, Taufique 28
Aral, Bahadir 74
Atassi, Sam 8
Atkinson, Chad 45
Auyuen, Wuong Jodi 53
Axelrod, Elizabeth 56
Azimaee, Mahmoud 41
B
Badisa, Ganesh 76
Baecke, Philippe 63
Baer, Darius 71
Bailey, Jeff 42
Balan, Tonya 32
Balinskaite, Violeta 64
Ball, Kathy 20
Bankston, Euwell 84
Barnes, Arila 45
Battaglia, Michael 17
Battiston, Christopher 22 , 31 , 57
Beatrous, Steve 6
Beaty, Brenda 13
Beaver, Allan 58
Beaver, James 7
Beaver, Richard 12
Becker, Matthew 51
Bedford, Denise 45
Bee, Brian 10
Bellara, Aarti 19 , 62
Bell, Bethany 23 , 2
Benjamin, William 37 , 56
Bennett, Donna 10 , 83
Bentley, John 8 , 46
Berryhill, Tim 47
Betsinger, Alicia 41
Bhardwaj, Pankaj 20
Bibb, Barbara 16
Bieringer, Alicia 59
Bigalke, Jan 83
Billings, Thomas 55 , 80 , 84
Birds, Andy 34
Bjurstrom, Jennifer 62
Bogard, Matt 7
Bolen, Tison 33
Bonham, Bob 35
Boniface, Christopher 7 , 54
Bonney, Christine 14
Boogaard, Sjoerd 78
Boorse, Carrie 50
Booth, Allison 57
Boscardin, John 33
Bost, Christopher 55 , 56
Bouedo, Mickael 6
Bretz, Michael 67
Bright, Terri 17
Brooks, Sandra 52
Brown, James R 54
Brown, Mark 71
Brown, Tony 67
Busi Reddy, Srinivas Reddy 24
C
Cai, Weijie 82
Carey, Delicia 23
Carlson, Rachel 78
Carpenter, Art 5 , 12 , 69
Carr, Brenda 58
Cathie, Andrew 37
Cegielski, Paul 7
Cenzer, Irena 33
Chadha, Rajbir 19 , 53
Chakraborty, Goutam 10 , 21 , 22 ,
23 , 24 , 25 , 28 , 37 , 38 , 53 , 71 , 76
Chandran, Saravana 31 , 83
Chapman, Don 7
Chen, Fang 2
Cheng, Alice 13
Cheng, Wei 78
Chen, Min 20
Chen, Qinghua (Kathy) 72
Chen, Wei 71
Chen, Yi-Hsin 18 , 19 , 62
Cheong, Michelle 9 , 21
Chew, Maureen 40
Chick, Brian 71
Chitale, Anand 41 , 61
Choy, Justin 32 , 70
Choy, Murphy 9 , 21 , 29
Christian, Stacey 47 , 71
Chung, Kevin 53
Chvosta, Jan 81
Clapson, Andrew 77
Clouse, Amy 59 , 79
Clover, Lina 61
Cochran, Ben 38
Cohen, Robert 80
Colavincenzo, Jenna 76
Coleman, Oita 7
Conley, Lisa 33
Conway, Ted 27
Corliss, David 40 , 83
Correa Bahnsen, Alejandro 43
Cox, James 59
Craig, Jean 15
Crain, Charlotte 42
Crawford, Peter 10 , 66
Crevar, Margaret 67
Cui, Xiangchen 20 , 75
Culley, Joan 15
Cunningham, John 8
Czika, Wendy 44
D
D'Agostino, Lucy 64
Darade, Harshal 27
Darden, Paul 26
Davies, Jennifer 30
Day, Eric 27
Day, Gavin 12 , 13
Day, Jeff 74
de Castro, Arnie 74
Deaton, Randall 56
Deguire, Yves 5
DelGobbo, Vince 48
Delwiche, Lora 47
Derby, Nate 52 , 54
Derr, Bob 65
deVille, Barry 45
Dhillon, Rupinder 12
Dickey, David 62
Dickinson, L. Miriam 13
Dingstad, Bernt 40
Ding, Sheng 26
Dong, Qunming 20
Dong, Tianxi 25
Donovan, Bill 14
Dorfman, Paul 47
Dowling, Molly 23
Droogendyk, Harry 10 , 72
Drutar, Michael 7
Dunn, Vann 14
Duraidhayalu, Hari harasudhan 24 ,
25
E
Ebben, Raymond 78
Eberhardt, Peter 6 , 11 , 19
Eckler, Lisa 22 , 30
Edwards, David 51
Elkin, Eric 13 , 2
Elliott, Alan 27
Ellis, Dylan 39
Ene, Mihaela 23 , 2
Erdman, Donald 47 , 71
Eubanks, Elizabeth 24
Ewing, Daphne 73
F
Faria, Plinio 41
Fecht, Marje 10 , 12
Fehd, Ronald 29 , 76
Ferguson, Ann 79
Flam Zalcman, Rosely 21
Flavin, Justina 13
Flinchum, Tamara 74
Florez Hormiga, Deybis 44
Fogleman, Stanley 28
Ford, Bill 12
Forst, Fred 66
Frost, Mike 42
Fu, I-kong 61
Fu, Yu 23 , 25 , 28 , 53
G
Gaethe, Gary 22
Gahagan, Ken 84
Galinson, Boaz 46
Gao, Nanxiang 43
Gao, Yubo 18
Gao, Yun 63
Gardiner, Joseph 3
Garla, Satish 37 , 51
Garrett, Guy 69
George, Tammi Kay 7
Ghadban, Khaled 12
Ghandehari, Heli 28 , 77
Ghanekar, Saurabh 25
Gibbs, Phil 3
Gibbs, Sandy 10
Gidley, Scott 42 , 43
Gilsen, Bruce 39 , 55
Glassman, Amy 70
Goldman, Robert 57
Gonzalez, Andres 43
Goodin, Shelly 14
Goodman, Melody S. 64
Gordon, Leonard 10
Greenfield, Jason 33
Gregory, Raymond 14
Greni, Chris 55
Gribbin, Dave 70
Grover, Swati 22
Gu, Haoyu 29
Guido, Lori 19 , 67
Gunes, Funda 1
Guo, Baojian 26
Gupta, Saurabh 59
Gupta, Sunil 1 , 30 , 73 , 80
Gutierrez, Bobby 33
H
Hadden, Louise 17
Haigh, Doug 66
Hale, Jessica 26
Hall, Angela 8 , 16 , 73
Haller, Susan 44
Hamilton, Jack 35 , 72
Hamm, Rob 35
Hamstra, Kirsten 14
Han, Yanwei 75
Hardin, J. Michael 32
Harris, Kriss 21
www.sasglobalforum.org/2013
85
Hastings, Valerie 77
Hatcher, Diane 35
Hatton, Robert 45
Hayes, Spencer 83
Heaton, John 36 , 42
Hebbar, Prashant 58
Hecht, Michael 37
Heiliger, Carsten 71
Heiney, Sue 65
Hembree, Tina 52
Hemedinger, Chris 5 , 61
Hendel, Russell 29
Henderson, Don 49
Henrick, Andrew 47
He, Tao 50
Hevener, Ginny 74
High, Robin 65
Hill, Aaron 60
Hill, Melissa 15
Hill, Toby 29
Hinson, Joseph 6
Hodges, Greg 7 , 80
Holland, Philip 69
Holman, James 7
Holmes, Harold 62
Holmes, Steven 40
Homes, Paul 66
Hopping, Albert 51
Horwitz, Lisa 52
Howell, Andrew 52
Hoyle, Larry 54
Huang, Chao 23
Huang, Tao 75
Huang, Zhongwen 17 , 50
Hughes, Ed 75
Hu, Jiannan 75
Hummel, Andrew 57
Hung, Dorothy 51
Hunley, Chuck 35
Hunter, Tim 69
Hurley, George 56
Hynan, Linda 27
Jiang, Hong 40
Jiang, Shaozong 43
Ji, Junyao 45
John, Christine 26
Johnson, Maribeth 1
Jones, Vicki 84
Jordan, Lori 70
Jordan, Mark 2
K
Ihnen, Leigh 84
Inbasekaran, Rajesh 25
Inoa, Ignacio A. 64
Islam, Mominul 75
Ito, Shunsuke 14
Izrael, David 17
Kachare, Atul 58
Kaduwela, Vijitha 25
Kakde, Deovrat 25
Kalgotra, Pankush 38
Kalidindi, Shiva 53
Karafa, Matt 6
Karimova, Mariya 26
Karnavat, Pranav 24
Kauhl, Justin 9
Keefer, Tom 34
Keener, Gordon 32
Keith, Jr., Michael 14
Kellermann, Anh 19
Kelley, David 57
Kelly, Patrick 56
Kenkare, Pragati 51
Kent, Paul 12 , 13 , 40 , 60
Keshan, Xia 39 , 56
Kezik, Julie 15
Khalil, Yehia 52
Khambhati, Shilpa 16
Khanna, Kamya 55
Kiernan, Kathleen 3
Kilburn, Julie 54
Kim, Eun Sook 19
Kim, Jonghyun 25
Kincaid, Chuck 48
King, Michael 35
Kirby, Katharine 33
Kong, Karena 77
Konya, Mark 20
Koval, Scott 69
Kratz, David 81
Kromrey, Jeffrey 18 , 19 , 20 , 62
Krycha, Karl 9
Kuhfeld, Warren 2 , 82
Kukhareva, Polina 1
Kuligowski, Andrew 49
Kunselman, Thomas 61
Kurtz, Joy 18
J
L
Jacob, Jeffin 22
Jansen, Lex 18
Javanainen, Jouni 66
Jia, Justin 78
Jiang, Charley 5
Jiang, Haibo 60
La Valley, Richard 18 , 45
Lada, Emily 74
Lafler, Kirk Paul 3 , 15 , 18 , 49 , 78
Lai, Ginny 13
Lan, Bob 15
Landy, David 18
I
86
www.sasglobalforum.org/2013
Langston, Julianna 57
Langston, Rick 39 , 76
Lau, Miu Ling 19
Lawhorn, Bari 69
Lee, Taiyeong 44
Lemieux, James 81
Lepkowski, James 5
Leslie, Elizabeth 27
Leslie, Scott 60
Lesser, Martin 26
Levine, Stuart 50
Lewis, Taylor 81
Li, Arthur 42 , 47
Li, Isaac 62
Li, Jian 43
Liming, Douglas 54
Lin, Alec 44
Lin, Amanda 78
Lin, Guixian 32
Link, Greg 74
Lin, Zhangxi 25 , 27
Li, Regan 15
Li, Siming 27
Li, Suwen 15
Liu, Charlie 30
LIU, FENG 61
Liu, Jiawen 22
Lix, Lisa 27
Li, Yahua 43
Li, Zhiyong 35
Lofland, Chelsea 77
Loman, Cynthia 15
Lopez, Adolfo 21
Lopez, Victor 28 , 77
Lou, Youbei 17
Low, Ronald 14
Ludwig, David 18
Lund, Bruce 39
Lund, Pete 39
Lu, Pei-Chin 82
Lycan, Christiana 8
M
Macaro, Christian 81
Macfarlane, Andrew 35
Maier, Mike 38
Maldonado, Miguel 44
Mandati, Siddhartha Reddy 21 , 76
Manickam, Airaha Chelvakkanthan
20 , 61
Mann, Robert 21
Maradiaga, David 77 , 82
Markey, Elaine 80
Massengill, Darrell 79
Ma, Sai 15
Matange, Sanjay 11 , 13 , 56 , 58
McCann, Claudia 43
McConnell, Lelia 79
McGahan, Colleen 54
McGaugh, Miriam 25 , 53
McGowan, Kevin 40
McLawhorn, Kathryn 72
McNeill, Bill 6
Meeran Mohideen, Musthan Kader
Ibrahim 22 , 24
Mehler, Gary 80
Mendelsohn, Andy 40
Mengelbier, Magnus 51
Meng, Xiangxiang 44
Miao, Yinghui 33
Miller, Tracie 18
Minjoe, Sandra 11 , 75
Miralles, Romain 55
Misra, Aditya 45
Mistler, Stephen 64
Mistry, Jugdish 31
Moeng, Koketso 78
Monaco, Rick 51
Moore, Stephen 19 , 67
Moreno-Simon, Tawney 28
Morton, Steve 69
Mu, George 13
Mullagiri, Sreenivas 80
Muller, Roger 25 , 26
Mulugeta, Dawit 33
Murphy, William 55
Myers, Keith 70
Myers, Susan 57
N
Na, Beeya 2
Nagarajan, Srihari 24
Nash, Michael 50
Nauta, Frank 58
Nelson, Gerry 83
Nelson, Greg 51 , 66
Nelson, Robert 23
Ngo, Theresa 82
Ng, Song Lin 22
Nguyen, Diep 19
Nguyen, Mai 16 , 24
Nisbet, Stuart 7
Nist, Pauline 13
Nizam, Azhar 30 , 57
Noga, Steve 15
Nori, Murali 70
O
O'Connor, Dan 38
O'Neil, Michael 43
Okerson, Barbara 17 , 31
Olson, Diane 72
Olson, Mike 12 , 13
Ong Yeru, Cally 9
Orange, Cary 71
Osborne, Anastasiya 30
Osborne, Mary 59
Ottesen, Rebecca 54 , 56 , 77
Otto, Greg 42
Overby Wilkerson, Amy 16
Overton, Stephen 8
P
Pakalapati, Tathabbai 20
Palaniappan, Latha 50
Pallone, Mark 5
Pan, Helen 65
Pan, Minghua 43
Pantangi, Anil Kumar 21
Parker, Chevell 38
Parks, Jennifer 66 , 67
Pass, Ray 73
Pasta, David 32
Pearson, Robert 82
Pease, Andrew 74
Pedersen, Casper 9
Peker, Ayesgul 74
Pelan, Margaret 58
Peravalli Venkata Naga, Krutharth
Kumar 72
Peters, Amy 35 , 65
Peterson, Jared 45
Petrova, Tatyana 42
Pham, Hung 7 , 54
Pham, Thanh 18
Pletcher, Rich 34
Plumley, Justin 59
Polak, Leonard 29
Pole, Greg 79
Ponnupandy, Menaga 81
Poppe, Frank 78
Potter, Ken 45
Poulsen, Rachel 65
Priest, Elisa 26
Prins, Jared 45
Pujula, Aude 77 , 82
Punuru, Janardhana 44
Purushothaman, Ramya 24 , 53 , 61
Puryear, Lindsey 74
Q
Qi, Lingxiao 43
Qin, Xiaojin 6
Qu, Hengrui 16
Qureshi, Lubna 51
R
Raimi, Steven 39
Raithel, Michael 1 , 46 , 66
Rajamani, Mythili 55
Ramalingam, Sanjiv 75
Rankin, Julie 79
Rasmussen, Patrice 18 , 62
Ratcliffe, Andrew 37 , 58
Rausch, Nancy 8 , 42 , 43
Ravuri, Sahithi 21
Rayabaram, Srikar 23 , 24 , 72
Ray, Robert 32
Reay, Stefanie 14
Redpath, Christopher 41
Rey, Timothy 9
Rhodes, Dianne Louise 52
Richardson, Brad 6
Richardson, Kari 11 , 12 , 73 , 74
Riddiough, Christine 1
Rigden, Chris 34
Rittman, Sarah 51
Rodriguez de Gil, Patricia 18 , 19 ,
20 , 62
Rodriguez-Deniz, Hector 74
Rodriguez, Bob 32 , 80
Roehl, William 5
Rogers, Stuart 67
Romano, Jeanine 18 , 62
Rosenbloom, Mary 5 , 78
Rossland, Eric 2 , 11 , 12 , 47 , 48 ,
49 , 73
Royal, Annette 41
Russell, Matthew 81
Ryan, Laura 44
S
Sadhukhan, Shreya 28
Sall, John 1
Sams, Scott 40
Sanders, Scott 58
Sanford, Kenneth 81
Sarkar, Deepa 55
Sarkar, Mantosh Kumar 10 , 22
Sauer, Brian 50
Schaan, Kathy 50
Schacherer, Chris 9 , 47 , 60
Schafer, Lori 58
Scheibmeir, John 83
Schmiedl, Ryan 46
Schmitz, Amber 51 , 60
Schneider, Frank 35
Schneider, Mark 83
Schoeneberger, Jason 2
Schreiber-Gregory, Deanna 82
Schuelke, Matthew 27
Scocca, David 52 , 75
Sedlak, Doug 32
Seffrin, Robert 31
Selvakumar, Prashanthi 77
Sempel, Hans 29
Sethi, Saratendu 45
Seth, Vivek 28
Setty, Ashok 59
Shah, Monarch 13
Shankar, Charu 49
Shao, Lucheng 46
Sharda, Ramesh 38
Sharma, Priya 59
Sharma, Ruchi 78
Shen, Dongmin 19
Shenvi, Neeta 30 , 57
Shigaev, Victor 29
Shim, Kyong Jin 29
Shin, John 72
Shipp, Charlie 15
Shive, Wanda 59
Shu, Amos 16 , 76
Shubert, David 40
Siddiqi, Naeem 44
Silva, Alan 18
Simon, Jim 2
Sindhu, Neetha 25
Singh, Sarwanjeet 53
Skillman, Shawn 70
Skoglund, Jimmy 71
Slaughter, Susan 47
Sloan, Stephen 74
Smiley, Whitney 23 , 2
Smith, Casey 81
Smith, Helen 24 , 57
Smith, Kevin 57
Song, Wen 55
Soto, Michael 71
So, Ying 2
Sperrazza, Casey 82
Springborn, Robert 38
Srivastava, Anurag 24
Stephens, Rob 83
Stokes, Maura 1
Styll, Rick 41 , 70
Suau-Sanchez, Pere 74
Surratt, Lane 44
Svendsen, Erik 15
Szeto, Nora 7 , 54
T
Tabachneck, Arthur 39 , 56
Taitel, Michael 17
Tan, Hui Fen 14
Tao, Jill 3
Tavakoli, Abbas 15 , 65
Terry, Robert 27
Thiyagarajan, Sreedevi 28
Thompson, Casey 35
Thompson, David 26
Thompson, Stephanie 44 , 73
Thompson, Wayne 9 , 32
Thornton, Patrick 31
Thota, Srikanth 20
Tilanus, Erik 6 , 53
Timusk, Peter 14
Tin Seong, Kam 45
Tobias, Randy 3
Trahan, Shane 24
Traubenberg, Seth 50
Trivedi, Bharat 8
V
Valliant, Richard 5
Valverde, Roberto 29
Van Daele, Douglas 22
Van den Poel, Dirk 63
Vanderlooy, Stijn 5
Vandervort, Eric 37
Varney, Brian 39
Venkataramani, Meera 80
Villiers, Peter 31
Virgile, Robert 39 , 56
Virji, Shirmeen 25 , 28 , 53
Vitron, Christine 48 , 50
Voltes-Dorta, Augusto 74
Vralstad, Svein Erik 42
W
Wachsmuth, Jason 54
Wagner, James 5
Waller, Jennifer 1 , 48
Walters, Barbara 84
Wang, Cindy 32
Wang, Lihui 21
Wang, Stacy 50
Wang, Xiyun (Cheryl) 5 , 37
Wang, Ying 43
Wang, Yongyin 72
Warman, Nicholson (Nick) 36
Watson, Greg 75
Weber, Tom 42
Weiss, Michael 37
Weiss, Mitchell 69
Wells, Chip 9
Wexler, Jonathan 9
Whitehurst, Joe 39 , 56
Wicklin, Rick 49
Widel, Mario 11 , 75
Wilkins, Scott 46
Williams, Christianna 3 , 48
Williams, Simon 35 , 67
Wilson, Michael G. 2
Wolfe, Bryan 65
Wong, Eric 50 , 51
Wu, Yi-Fang 27
X
Xie, Fagen 43
Xu, Meili 18
Xu, Yan 75
Y
Yang, Dongsheng 24
Yang, Jason 55
Yao, Xue 27
Yiu, Sau 54
Yong, Chin Khian 22
Yuan, Yang 33
www.sasglobalforum.org/2013
87
Z
Zakzeski, Audra 31
Zapata, Hector 82
Zaratsian, Dan 59
Zender, Cynthia 46 , 57
Zhang, Jiawen 43
Zhang, Ruiwen 44 , 61
Zhang, Sijian 76
Zhang, Xin 30 , 57
Zhang, Yu 43
Zhao, Beinan 50
Zhao, Zheng 59
zhao, juan 16
Zhong, Christina 75
Zink, Richard 75
Zuniga, Daniel 34
Zupko, William 52
88
www.sasglobalforum.org/2013