chief data scientist Archives | FedScoop https://fedscoop.com/tag/chief-data-scientist/ FedScoop delivers up-to-the-minute breaking government tech news and is the government IT community's platform for education and collaboration through news, events, radio and TV. FedScoop engages top leaders from the White House, federal agencies, academia and the tech industry both online and in person to discuss ways technology can improve government, and to exchange best practices and identify how to achieve common goals. Fri, 09 Sep 2022 21:05:09 +0000 en-US hourly 1 https://wordpress.org/?v=6.5.4 https://fedscoop.com/wp-content/uploads/sites/5/2023/01/cropped-fs_favicon-3.png?w=32 chief data scientist Archives | FedScoop https://fedscoop.com/tag/chief-data-scientist/ 32 32 U.S. chief data scientist explains government’s push for greater use of disaggregated data https://fedscoop.com/us-chief-data-scientist-interview/ Fri, 09 Sep 2022 19:52:59 +0000 https://fedscoop.com/?p=60032 Denice Ross spells out her priorities over the last 10 months leading the government's data strategy.

The post U.S. chief data scientist explains government’s push for greater use of disaggregated data appeared first on FedScoop.

]]>
U.S. Chief Data Scientist Denice Ross remembers the killing of 18-year-old Michael Brown by Ferguson, Missouri police as a galvanizing moment for federal officials in their approach to open data.

At the time of the young man’s death in 2014, police departments did not release use of force data — basic information needed by federal officials to determine how Black communities were being affected by law enforcement violence — and led her to spearhead the novel Police Data Initiative.

The effort started with 14 police departments committing to opening at least three datasets — their use of force dataset almost always being one — and 129 jurisdictions were on board by the end of the Obama administration. Police and citizen expectations of what transparency and accountability should look like, and what data should be open, had changed, Ross told FedScoop in an exclusive interview.

“But we didn’t create a mechanism for turning that data into action, so that’s why I’m back.” Ross said. “Because open data is necessary and not sufficient to drive the type of action that we need to create a more equitable society.”

Ross was a Presidential Innovation Fellow at the time.

Now the U.S. chief data scientist since November, Ross’ focus has been ensuring the data agencies are using and publishing are yielding more equitable outcomes for Americans. And that requires the “next generation of open data” as she sees it: disaggregated data.

Disaggregated data is separated into smaller units, often demographically, to answer questions like which populations are underserved by federal programs and policies and make course corrections narrowing service gaps. The process is time intensive and necessitates skilled data practitioners, including career federal officials upskilled in data science, Ross said.

President Biden stressed his commitment to equitable data from his first day in office with the immediate issuance of the Racial Equity Executive Order in support of underserved communities. When the White House’s Equitable Data Working Group (EDWG), created by the order, released its five recommendations in April for improving use, normalizing disaggregated data while protecting privacy topped the list.

“Now everybody is talking about disaggregating data,” Ross said. “It used to be — I’ve been in this field for 20 years — I always avoided that word because it was so jargoney, and nobody knew what we were talking about.”

Still agencies are a “mixed bag” when it comes to collecting disaggregated data and using it properly, she added.

Bright spots include the Federal Interagency Council on Statistical Policy recently releasing a searchable catalog of disaggregated datasets on Asian, Native Hawaiian and Pacific Islander populations, as well as agencies disaggregating grant data by location to ensure fairer distribution, Ross said.

The chief data scientist has spent the last 10 months building a small team within the White House Office of Science and Technology Policy that supports the Biden administration’s biggest priorities like the Bipartisan Infrastructure Law (BIL) with $1.2 billion behind it, Inflation Reduction Act and Customer Experience Executive Order reducing barriers to government benefits. For the first time, a dedicated team is applying disaggregated data to a president’s policy agenda.

“What I do is infuse equitable data into those priorities, so data isn’t a side thing,” Ross said. “It’s actually integrated into how we design programs and policies.”

For that reason it’s important the team be a diverse mix of genders, races, ethnicities and lived experiences, she added.

Ross finds the biggest obstacle to the team’s work is its hybrid nature; there aren’t as many in-person interagency meetings or civic tech innovation summits for sharing best practices since the pandemic began.

“The collaboration tools that we have are just mostly not compatible,” Ross said. “And so we end up making the most of what we can with a PowerPoint and a Zoom call, but that’s a far cry from being in the same room with a bunch of post-it notes and really doing some solid design thinking using the best available tools.”

Ross assists with the hiring of data practitioners within agencies like the Office of Personnel Management’s surge team for the White House’s BIL implementation, which requires hundreds of STEM-trained personnel to support investments.

The chief data scientist’s team is responsible for operationalizing some of the recommendations of the White House Equitable Data Working Group (EDWG), which is transitioning into the National Science and Technology Council’s Subcommittee on Equitable Data. Ross will co-chair that subcommittee.

In addition to disaggregating data, the team is uncovering underused data; improving agencies’ capacity for policy and program equity assessments; creating public data visualization tools; and soliciting feedback from state, local, tribal and territorial communities. Some communities surpass the federal government when it comes to disaggregating data, which is why OSTP recently issued requests for information (RFIs) on LGBTQI+ equity and equitable data engagement and accountability on behalf of the new subcommittee.

“We’re really serious about these RFIs because we need the wisdom from the field in order to be able to implement these Equitable Data Working Group recommendations, in the most useful way, inside the federal government,” Ross said.

Other EDWG recommendations fall to the Office of Management and Budget and U.S. Chief Statistician Karin Orvis, who’s currently modernizing the 25-year-old Statistical Policy Directive No. 15 (SPD 15) on race and ethnicity data standards. 

OMB recently released a plain-language recommendation to agencies for making the best use of existing race and ethnicity standards because the revised guidance isn’t expected until summer 2024. The recommendation includes practical flexibilities for disaggregating race and ethnicity data, approaches to data on more than one race, and advice on adding additional race categories to forms or surveys.

“I’ll just spoil that one,” Ross said. “You should not add some other race category to your forms or surveys because then it makes your data really unusable.”

OMB’s first listening session on the SPD 15 revision is slated for Sept. 15, 2022, and an RFI will be issued soon, Ross said.

The chief data scientist expects the update will have a ripple effect on how SLTT governments collect their own data with California already considering new race categories.

Ross will spend the rest of the year helping stand up the Working Group on Criminal Justice Statistics called for in May’s Policing and Criminal Justice Executive Order, harkening back to her work as a PIF, and ensuring her subcommittee hits the ground running.

“My priority for the rest of 2022 is to get these interagency collaborations going through the Subcommittee on Equitable Data,” Ross said. “That includes working on sexual orientation and gender identity data, infrastructure investment and equity assessments.”

The post U.S. chief data scientist explains government’s push for greater use of disaggregated data appeared first on FedScoop.

]]>
60032
GAO says AI oversight framework will help in continuously monitoring agencies https://fedscoop.com/ai-oversight-gao-framework/ https://fedscoop.com/ai-oversight-gao-framework/#respond Wed, 11 Nov 2020 21:19:50 +0000 https://fedscoop.com/?p=38853 Federal officials have floated the possibility of the AI framework functioning like a scorecard for agencies. The GSA's Taka Ariga says it needs "practical" principles if it's going to work.

The post GAO says AI oversight framework will help in continuously monitoring agencies appeared first on FedScoop.

]]>
The Government Accountability Office’s forthcoming artificial intelligence oversight framework will help auditors work with inspectors general to continuously monitor executive agencies’ progress with the technology, says the agency’s first-ever chief data scientist.

Right now GAO is developing a “wireframe” of what oversight might look like in the areas of explainability and transparency, bias and fairness, integrity and resilience, and data quality and lineage, said Taka Ariga, who also directs his agency’s Innovation Lab.

Federal officials floated the possibility of the AI framework functioning like a scorecard for agencies in September, and Ariga said “practical” principles are needed beyond simply “do no harm.”

“The proliferation of data has really accelerated machine learning deployment,” Ariga said, during a Data Coalition event Tuesday. “Coming with that are a lot of disparate impacts, societal impacts that at this point are very high level from a governance perspective.”

GAO stands to benefit as well as it employs advanced analytics on other agencies’ open source, subscription, operational, and classified data in the course of its audits, he added.

The agency is conducting more than 100 pandemic-related audits and evaluations on top of its usual oversight and has aspirations of prototyping AI, cloud services and blockchain to help securely process all the incoming data, Ariga said.

At the same time the Innovation Lab is fostering holistic data literacy among GAO auditors.

“Complicated analytical exercises shouldn’t always be outsourced to data scientists, statisticians and other specialized skillsets,” Ariga said. “We want all GAO personnel to have generalized awareness of what ‘right’ looks like and what ‘wrong’ looks like.”

GAO is implementing a data catalog so auditors can see what information is available and how they can use it, which also requires policies for who can access what.

And pipelines and containers are being developed for algorithms so they can not only scale but have their code “tweaked” as needed based on changes in requirements and circumstances, rather than having to be refactored or entirely redeveloped, Ariga said.

The Innovation Lab is taking care of data transfers, quality and visualization so GAO analysts can focus on researchable questions using the information.

“I consider all of those utility functions just like electricity, just like water,” Ariga said. “When you turn them on, they should just work.”

The post GAO says AI oversight framework will help in continuously monitoring agencies appeared first on FedScoop.

]]>
https://fedscoop.com/ai-oversight-gao-framework/feed/ 0 38853