“The Death Master File (DMF) is a computer database file made available by the United States Social Security Administration since 1980″ according to Wikipedia. It is available here https://ladmf.ntis.gov/ but costs $2,930.00 anually.
The file has since been posted on the internet for free, including here:
This file can be useful for OSINT (Open Source Intelligence) discovery as it contains millions of US SSNs (Social Security Numbers), so we have decided to add it to our search index.
Here is the file indexed in our search engine: https://intelx.io/?did=fd36a1b3-35ff-429b-8c19-e6a4e229ffb9
The Death Master File is a text file with 100 characters per line each representing one record. This is the official format specification according to https://dmf.ntis.gov/recordlayout.pdf:
Sample records of the actual file:
001010001MUZZEY GRACE 1200197504161902 001010009SMITH ROGER 0400196902041892 001010010HAMMOND KENNETH 0300197604241904 001010011DREW LEON R V0830198706141908
Converting the file from its proprietary format into CSV is necessary for for proper indexing by our search engine and for other uses cases such as opening it manually in Excel (setting aside the size of the file).
We have published the code as open source here: https://github.com/IntelligenceX/DeathMasterFile2CSV
This is the result of the above records in CSV:
Type,Social Security Number,Last Name,Name Suffix,First Name,Middle Name,Verified,Date of Death,Date of Birth,Blank 1,Blank2,Blank 3,Blank 4 ,001010001,Muzzey,,Grace,,,1975-12-00,1902-04-16,,,, ,001010009,Smith,,Roger,,,1969-04-00,1892-02-04,,,, ,001010010,Hammond,,Kenneth,,,1976-03-00,1904-04-24,,,, ,001010011,Drew,,Leon,R,Verified,1987-08-30,1908-06-14,,,,
One of the few things we had to decide was the date format. Internally we are always using “YYYY-MM-DD” and googling for the answer, confirmed our choice, according to Stackoverflow:
Another choice we made was to camel-case the names (“MUZZEY GRACE” becomes “Muzzey Grace”) to improve human readability.
The converter reads 100 MB chunks at a time and processes it, as the input file (2011 version) is 8 GB of size. It took 5 minutes to convert its 85,822,194 records to CSV. The resulting CSV file is only 4.6 GB of size, since unnecessary white spaces are removed.
The 2013 version has 87,735,016 records and is 8.3 GB big. The converted CSV file is 4.7 GB big.
March 2020 Updates How we prepared for the current global situation We have ordered and deployed 180 TB worth of enterprise storage to be prepared for upcoming price increases and shortages of hard disks. We are reading reports that warn of upcoming delivery delays due to disruptions in the supply chain. We tweeted a picture
On December 11, 2019 we have received 31,866 HTTP requests from the IP 220.127.116.11. Below are few sample log entries: 18.104.22.168 – – [11/Dec/2019:13:52:37 +0000] “POST /login HTTP/1.1” 200 7448 22.214.171.124 – – [11/Dec/2019:13:52:37 +0000] “POST /login HTTP/1.1” 200 7443 126.96.36.199 – – [11/Dec/2019:13:52:37 +0000] “POST /login HTTP/1.1” 200 7447 188.8.131.52 – – [11/Dec/2019:13:52:37 +0000]
March 2020: New Data Categories Stay up to date with us on Twitter: https://twitter.com/_IntelligenceX Private Data Leaks We have revised our license system and launched a new data category exclusively for paid Professional users: Private Data Leaks. Trial accounts will see a preview, but you must be a paid member to fully access the category.