EconPapers    
Economics at your fingertips  
 

Privacy Guarantees in Synthetic Data

Sergey I. Nikolenko ()
Additional contact information
Sergey I. Nikolenko: Synthesis AI

Chapter Chapter 11 in Synthetic Data for Deep Learning, 2021, pp 269-283 from Springer

Abstract: Abstract In this chapter, we discuss another important field of applications for synthetic data: ensuring privacy. In many real-world problems, real data is sensitive enough that it is impossible to release. One possible solution could be to train generative models that would produce new synthetic datasets based on real data, while the real data itself would remain secret. But how can we be sure that real data will not be inadvertently leaked? Guarantees in this regard can be provided by the framework of differential privacy. We give a brief introduction to differential privacy, its relation to machine learning, and the guarantees that it can provide for synthetic data generation.

Date: 2021
References: Add references at CitEc
Citations:

There are no downloads for this item, see the EconPapers FAQ for hints about obtaining it.

Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.

Export reference: BibTeX RIS (EndNote, ProCite, RefMan) HTML/Text

Persistent link: https://EconPapers.repec.org/RePEc:spr:spochp:978-3-030-75178-4_11

Ordering information: This item can be ordered from
http://www.springer.com/9783030751784

DOI: 10.1007/978-3-030-75178-4_11

Access Statistics for this chapter

More chapters in Springer Optimization and Its Applications from Springer
Bibliographic data for series maintained by Sonal Shukla () and Springer Nature Abstracting and Indexing ().

 
Page updated 2025-06-06
Handle: RePEc:spr:spochp:978-3-030-75178-4_11