Page 1 of 2 12 LastLast
Results 1 to 16 of 17

Thread: Which file system for enormous number of files in one directory?

  1. #1
    Senior Member
    Join Date
    Jul 2003
    Location
    Reading, Berkshire
    Posts
    1,253
    Thanks
    64
    Thanked
    53 times in 34 posts
    • tfboy's system
      • Motherboard:
      • MSI X470 Gaming Plus
      • CPU:
      • AMD Ryzen 7 2700
      • Memory:
      • 2x8GB Corsair Vengeance LPX)
      • Storage:
      • Force MP600 1TB PCIe SSD
      • Graphics card(s):
      • 560 Ti
      • PSU:
      • Corsair RM 650W
      • Case:
      • CM Silencio 550
      • Operating System:
      • W10 Pro
      • Monitor(s):
      • HP LP2475w + Dell 2001FP
      • Internet:
      • VM 350Mb

    Which file system for enormous number of files in one directory?

    I'm working on creating a Digital Cinema Package for use with some professional projectors.

    The creation involves having each frame of video as a single JPEG2000 file. This is then muxed into an MXF stream and compiled into a DCP package.

    The issue I'm having at the moment is terrible slowdowns using NTFS on my server. I've already checked the volume's free space and fragmentation, and they're not that bad. Yet copying / moving files is horrendously slow - around one 1MB file per second.

    I reckon it's due to having all the files in a single directory. Unfortunately, I can't really split it up because the muxer needs all the files in one directory which composes a video stream.

    So what options can you suggest?
    - Would symbolic links work? I've never tried this in Windows / NTFS and have limited experience using ext3/4
    - Should I switch to another file system? If so, which one supports 200,000+ files in one directory and is usable on Windows and is free? I would use StorNext and it would work but it's hardly an option with the hardware I have (no SAN, metadata controllers, etc).

    Thoughts?

  2. #2
    Gentoo Ricer
    Join Date
    Jan 2005
    Location
    Galway
    Posts
    11,048
    Thanks
    1,016
    Thanked
    944 times in 704 posts
    • aidanjt's system
      • Motherboard:
      • Asus Strix Z370-G
      • CPU:
      • Intel i7-8700K
      • Memory:
      • 2x8GB Corsiar LPX 3000C15
      • Storage:
      • 500GB Samsung 960 EVO
      • Graphics card(s):
      • EVGA GTX 970 SC ACX 2.0
      • PSU:
      • EVGA G3 750W
      • Case:
      • Fractal Design Define C Mini
      • Operating System:
      • Windows 10 Pro
      • Monitor(s):
      • Asus MG279Q
      • Internet:
      • 240mbps Virgin Cable

    Re: Which file system for enormous number of files in one directory?

    With Windows, you only *really* have two choices, FAT32 or NTFS, and FAT32 is far less capable and not any faster.

    Perhaps you could suggest to the software vendor to add directory traversal support.
    Quote Originally Posted by Agent View Post
    ...every time Creative bring out a new card range their advertising makes it sound like they have discovered a way to insert a thousand Chuck Norris super dwarfs in your ears...

  3. #3
    Goron goron Kumagoro's Avatar
    Join Date
    Mar 2004
    Posts
    3,147
    Thanks
    37
    Thanked
    170 times in 139 posts

    Re: Which file system for enormous number of files in one directory?

    I have no idea if it will make any difference but there is also exfat
    Would indexing the folder make any difference?

  4. #4
    The late but legendary peterb - Onward and Upward peterb's Avatar
    Join Date
    Aug 2005
    Location
    Looking down & checking on swearing
    Posts
    19,378
    Thanks
    2,892
    Thanked
    3,403 times in 2,693 posts

    Re: Which file system for enormous number of files in one directory?

    I think Win 7 indexes folders by default, but that is for speed of searching rather than file transfer speed.
    (\__/)
    (='.'=)
    (")_(")

    Been helped or just 'Like' a post? Use the Thanks button!
    My broadband speed - 750 Meganibbles/minute

  5. #5
    DILLIGAF GoNz0's Avatar
    Join Date
    Jun 2006
    Location
    Derby
    Posts
    10,872
    Thanks
    632
    Thanked
    1,192 times in 945 posts
    • GoNz0's system
      • Motherboard:
      • Asus Rampage V Extreme
      • CPU:
      • i7 something X99 based
      • Memory:
      • 16gb GSkill
      • Storage:
      • 4 SSD's + WD Red
      • Graphics card(s):
      • GTX980 Strix WC
      • PSU:
      • Enermax Galaxy 1250 (9 years and counting)
      • Case:
      • Corsair 900D
      • Operating System:
      • win10 64bit
      • Monitor(s):
      • Dell 24"
      • Internet:
      • 220mb Cable

    Re: Which file system for enormous number of files in one directory?

    is this just for moving files about or is the software grinding to a halt processing so many jpegs ?

    if its just moving stuff back and forth ready for the muxing use richcopy

    Capitalization is the difference between helping your Uncle Jack
    off a horse and helping your uncle jack off a horse.

  6. #6
    Senior Member Betty_Swallocks's Avatar
    Join Date
    Jan 2005
    Location
    Feet up, spliff lit.
    Posts
    1,140
    Thanks
    70
    Thanked
    60 times in 44 posts
    • Betty_Swallocks's system
      • Motherboard:
      • Asus Z97-A
      • CPU:
      • Intel Core i5 4690K o/c to 4.6 gHz
      • Memory:
      • 8Gb DDR3
      • Storage:
      • 256Gb SSD + 1320Gb (3x SATA drives)
      • Graphics card(s):
      • MSI R9 390 8Gb
      • PSU:
      • Corsair CS750M
      • Case:
      • Thermaltake Shark
      • Operating System:
      • Windows 10
      • Monitor(s):
      • 37" Samsung TV @1920x1080 + Dell 20.1" TFT secondary screen
      • Internet:
      • 150Mb Virgin Media cable

    Re: Which file system for enormous number of files in one directory?

    You might find Linux handles it better.
    "Free speech includes not only the inoffensive but the irritating, the contentious, the eccentric, the heretical, the unwelcome and the provocative provided it does not tend to provoke violence. Freedom only to speak inoffensively is not worth having."

  7. #7
    The late but legendary peterb - Onward and Upward peterb's Avatar
    Join Date
    Aug 2005
    Location
    Looking down & checking on swearing
    Posts
    19,378
    Thanks
    2,892
    Thanked
    3,403 times in 2,693 posts

    Re: Which file system for enormous number of files in one directory?

    Quote Originally Posted by Betty_Swallocks View Post
    You might find Linux handles it better.
    There are certainly more filesystems available under a linux kernel, and xfs (or jfs) would probably be better for large numbers of large files. Of those, XFS is probably the better one because although it is slightly slower in tests, it is less processor intensive, which might be important if you are doing image manipulation at the same time.

    Of course that is all academic unless it is the storage that is the problem (and you can store on a separate linux machine) or the tools that you are using (or ones with similar functionality) are also available on a linux platform.

    Some links you might find useful:

    http://www.debian-administration.org/articles/388

    (xfs is available for other distros, it isn't debian specific)

    http://xfs.org/index.php/XFS_FAQ
    (\__/)
    (='.'=)
    (")_(")

    Been helped or just 'Like' a post? Use the Thanks button!
    My broadband speed - 750 Meganibbles/minute

  8. #8
    Member
    Join Date
    Aug 2006
    Location
    London
    Posts
    188
    Thanks
    12
    Thanked
    7 times in 6 posts
    • Savas's system
      • Motherboard:
      • Asus P6X58D-E
      • CPU:
      • Intel Core i7 950 @ 3.07 GHz
      • Memory:
      • 6GB Corsair
      • Storage:
      • SanDisk SSD PLUS 240 GB Sata III 2.5 inch Internal SSD
      • Graphics card(s):
      • Gigabyte GeForce GTX 460
      • PSU:
      • Corsair HX850W
      • Case:
      • LIANLI PC-8FI
      • Operating System:
      • Windows 10
      • Monitor(s):
      • LG 32LF580V
      • Internet:
      • 80Mbits (Plusnet)

    Re: Which file system for enormous number of files in one directory?

    What operating systems are you using?

    If there's a lot of files in one directory, then maybe something to consider is directory caching, so you won't constantly be retrieving each and every single file information every time it's in the dir. Basically things like offline mode will be your interests.

    http://trycatch.be/blogs/roggenk/arc...t-caching.aspx

    I guess there's something for linux too.

  9. #9
    Not a good person scaryjim's Avatar
    Join Date
    Jan 2009
    Location
    Gateshead
    Posts
    15,196
    Thanks
    1,231
    Thanked
    2,291 times in 1,874 posts
    • scaryjim's system
      • Motherboard:
      • Dell Inspiron
      • CPU:
      • Core i5 8250U
      • Memory:
      • 2x 4GB DDR4 2666
      • Storage:
      • 128GB M.2 SSD + 1TB HDD
      • Graphics card(s):
      • Radeon R5 230
      • PSU:
      • Battery/Dell brick
      • Case:
      • Dell Inspiron 5570
      • Operating System:
      • Windows 10
      • Monitor(s):
      • 15" 1080p laptop panel

    Re: Which file system for enormous number of files in one directory?

    Just a random guess, but I'd assume the issue is that you're asking the program to do a lot of random reads, and you're using a mechnical disk: might it be worth shunting the whole lot onto an SSD and seeing if it speeds up? If you're talking vast numbers of 1MB files then, even if the filesystems thinks they're all in one directory, chances are they're scattered all over the physical disk, and since you'll need to access them in a specific order you could be looking at a lot of worse-case seek times.

    Obviously the ideal situation would be having a spare SSD you could test this theory on first, but a small (i.e. 30GB) SSD wouldn't be that expensive an experiment...

    EDIT: Ebuyer have a 60GB Vertex Plus for £58...

  10. #10
    Senior Member
    Join Date
    Jul 2003
    Location
    Reading, Berkshire
    Posts
    1,253
    Thanks
    64
    Thanked
    53 times in 34 posts
    • tfboy's system
      • Motherboard:
      • MSI X470 Gaming Plus
      • CPU:
      • AMD Ryzen 7 2700
      • Memory:
      • 2x8GB Corsair Vengeance LPX)
      • Storage:
      • Force MP600 1TB PCIe SSD
      • Graphics card(s):
      • 560 Ti
      • PSU:
      • Corsair RM 650W
      • Case:
      • CM Silencio 550
      • Operating System:
      • W10 Pro
      • Monitor(s):
      • HP LP2475w + Dell 2001FP
      • Internet:
      • VM 350Mb

    Re: Which file system for enormous number of files in one directory?

    The storage server is a Windows Home Server (v1), with some 2TB drives - 5400rpm slow ones!.

    I'll quickly explain the workflow as there are a few stages to it: I'm creating a 3D DCP, and that involves having individual left and right eye stereoscopic files in jpeg2000 format, XYZ colour space. I'm using an open source program for the j2k and xyz conversion (OpenDCP). This is cross-platform and I could have it running on Linux. However, it's the storage that's the problem.

    1) The single large video file is parsed through Adobe Media Encoder to create individual TIFF stills of each frame of the video. These frames are cropped and resized on the fly by AME. Unfortunately, the TIFF export of AME is uncompressed TIFF only, so you're looking at 8MB per TIFF file. At 24 files per second, then multiply by two to have your left and right sets of files, you can see how you quickly need several terabytes

    2) The resulting TIFF is then read by OpenDCP to create a the j2k/xyz version. The actual executable of OpenDCP is run on several workstations so I can have a kind of render farm - I'm moving groups of TIFF files into individual directories to farm out the load, typically 10,000 files per directory.

    3) OpenDCP saves the resulting j2k file somewhere else - it's now on a different physical drive, but still inside the same server.

    I now realise one of the potential problems is that I was using a disk in the storage pool which probably has an adverse affect on performance.

    In terms of random read / write, I suppose it is. I have, at most, 12 simultaneous accesses from remote machines (mapped drives) to read the files, compress them and write back the Jpeg2000 versions. I've now started writing the jp2k ones on a different physical drive (still inside the server), just in case it was a read/write issue.

    Since writing the original post, I've realised that there had been massive fragmentation as the result of writing the initial files. It's taken around 6 hours to half defragment! Having defragmented, it's slightly faster, but still slow, so I think it might be more an NTFS limitation than a physical fragmentation problem.

    I do have an EXT3 (maybe 4) Windows driver somewhere that works fine. How the performance is though, I don't know. I guess I'd be better with a small Linux NAS storage box to hold the data, as I can't afford to wipe my existing WHS installation. Thanks for the suggestion Peter

    Jim, I'd love to use SSDs, however, 2TB+ worth isn't going to be cheap

  11. #11
    DILLIGAF GoNz0's Avatar
    Join Date
    Jun 2006
    Location
    Derby
    Posts
    10,872
    Thanks
    632
    Thanked
    1,192 times in 945 posts
    • GoNz0's system
      • Motherboard:
      • Asus Rampage V Extreme
      • CPU:
      • i7 something X99 based
      • Memory:
      • 16gb GSkill
      • Storage:
      • 4 SSD's + WD Red
      • Graphics card(s):
      • GTX980 Strix WC
      • PSU:
      • Enermax Galaxy 1250 (9 years and counting)
      • Case:
      • Corsair 900D
      • Operating System:
      • win10 64bit
      • Monitor(s):
      • Dell 24"
      • Internet:
      • 220mb Cable

    Re: Which file system for enormous number of files in one directory?

    no wonder its slow, less than ideal situation m8.

    Capitalization is the difference between helping your Uncle Jack
    off a horse and helping your uncle jack off a horse.

  12. #12
    Senior Member
    Join Date
    Jul 2003
    Location
    Reading, Berkshire
    Posts
    1,253
    Thanks
    64
    Thanked
    53 times in 34 posts
    • tfboy's system
      • Motherboard:
      • MSI X470 Gaming Plus
      • CPU:
      • AMD Ryzen 7 2700
      • Memory:
      • 2x8GB Corsair Vengeance LPX)
      • Storage:
      • Force MP600 1TB PCIe SSD
      • Graphics card(s):
      • 560 Ti
      • PSU:
      • Corsair RM 650W
      • Case:
      • CM Silencio 550
      • Operating System:
      • W10 Pro
      • Monitor(s):
      • HP LP2475w + Dell 2001FP
      • Internet:
      • VM 350Mb

    Re: Which file system for enormous number of files in one directory?

    LOL Gonzo. Well, I'd love to have a proper DPX post prod film workstation with a zippy SAN behind it, but it's a bit beyond my budget! I suppose I ought to try and source some hardware from work to do it as it's half a work test project (compiling 3D stereoscopic DCPs).

  13. #13
    The late but legendary peterb - Onward and Upward peterb's Avatar
    Join Date
    Aug 2005
    Location
    Looking down & checking on swearing
    Posts
    19,378
    Thanks
    2,892
    Thanked
    3,403 times in 2,693 posts

    Re: Which file system for enormous number of files in one directory?

    It does sound as if the set up is underpowered! If your budget can stretch to it, something like a Mac Pro with multiple processors (I have seen one with 4 six core processors for example) would make the whole process a lot quicker - but you need to check t hat the software you are using can take optimum advantage of it - but I guess if you are doing s project of that magnitude, you are pretty well up on software capabilities and limitations.
    (\__/)
    (='.'=)
    (")_(")

    Been helped or just 'Like' a post? Use the Thanks button!
    My broadband speed - 750 Meganibbles/minute

  14. #14
    Senior Member
    Join Date
    Jul 2003
    Location
    Reading, Berkshire
    Posts
    1,253
    Thanks
    64
    Thanked
    53 times in 34 posts
    • tfboy's system
      • Motherboard:
      • MSI X470 Gaming Plus
      • CPU:
      • AMD Ryzen 7 2700
      • Memory:
      • 2x8GB Corsair Vengeance LPX)
      • Storage:
      • Force MP600 1TB PCIe SSD
      • Graphics card(s):
      • 560 Ti
      • PSU:
      • Corsair RM 650W
      • Case:
      • CM Silencio 550
      • Operating System:
      • W10 Pro
      • Monitor(s):
      • HP LP2475w + Dell 2001FP
      • Internet:
      • VM 350Mb

    Re: Which file system for enormous number of files in one directory?

    The processing power of recompressing into jpeg2000 isn't the problem, it's the storage subsystem. Interestingly, I just bought a new laptop with 2.8GHz i7 and it's faster than my desktop Q6600 running at 3GHz when it comes to this compression stuff... Anyway, I digress...

    Open DCP is multithreaded, but it's not that smart - it only parallel processes as many simultaneous files as you have threads, so having one instance using 4 threads is no different than having 4 instances only using 1 thread. Having faster processors would allow me to compress faster, but I don't think that's the bottleneck, it's the storage and file system that starts to struggle with the hundreds of thousands of files.

    I've just downloaded and installed PerfectDisk 12 defragmenter to see how well it works - it claims it reduces fragmentation by correctly allocating space before the files are written. I'll see how well it achieves what it claims very quickly.

  15. #15
    Does he need a reason? Funkstar's Avatar
    Join Date
    Aug 2005
    Location
    Aberdeen
    Posts
    19,874
    Thanks
    630
    Thanked
    965 times in 816 posts
    • Funkstar's system
      • Motherboard:
      • Gigabyte EG45M-DS2H
      • CPU:
      • Intel Core2Quad Q9550 (2.83GHz)
      • Memory:
      • 8GB OCZ PC2-6400C5 800MHz Quad Channel
      • Storage:
      • 650GB Western Digital Caviar Blue
      • Graphics card(s):
      • 512MB ATI Radeon HD4550
      • PSU:
      • Antec 350W 80+ Efficient PSU
      • Case:
      • Antec NSK1480 Slim Mini Desktop Case
      • Operating System:
      • Vista Ultimate 64bit
      • Monitor(s):
      • Dell 2407 + 2408 monitors
      • Internet:
      • Zen 8mb

    Re: Which file system for enormous number of files in one directory?

    I wonder if something like a PERC RAID card with those drives and then just sharing the drive over your network would be better than using a WHS box. I can't help feeling that real hardware RAID (even one on a budget) would be better than software sort-of-but-not-quite-RAID on Home Server.

  16. #16
    Senior Member
    Join Date
    Mar 2005
    Posts
    4,935
    Thanks
    171
    Thanked
    384 times in 311 posts
    • badass's system
      • Motherboard:
      • ASUS P8Z77-m pro
      • CPU:
      • Core i5 3570K
      • Memory:
      • 32GB
      • Storage:
      • 1TB Samsung 850 EVO, 2TB WD Green
      • Graphics card(s):
      • Radeon RX 580
      • PSU:
      • Corsair HX520W
      • Case:
      • Silverstone SG02-F
      • Operating System:
      • Windows 10 X64
      • Monitor(s):
      • Del U2311, LG226WTQ
      • Internet:
      • 80/20 FTTC

    Re: Which file system for enormous number of files in one directory?

    Quote Originally Posted by tfboy View Post
    The storage server is a Windows Home Server (v1), with some 2TB drives - 5400rpm slow ones!.

    I'll quickly explain the workflow as there are a few stages to it: I'm creating a 3D DCP, and that involves having individual left and right eye stereoscopic files in jpeg2000 format, XYZ colour space. I'm using an open source program for the j2k and xyz conversion (OpenDCP). This is cross-platform and I could have it running on Linux. However, it's the storage that's the problem.

    1) The single large video file is parsed through Adobe Media Encoder to create individual TIFF stills of each frame of the video. These frames are cropped and resized on the fly by AME. Unfortunately, the TIFF export of AME is uncompressed TIFF only, so you're looking at 8MB per TIFF file. At 24 files per second, then multiply by two to have your left and right sets of files, you can see how you quickly need several terabytes

    2) The resulting TIFF is then read by OpenDCP to create a the j2k/xyz version. The actual executable of OpenDCP is run on several workstations so I can have a kind of render farm - I'm moving groups of TIFF files into individual directories to farm out the load, typically 10,000 files per directory.

    3) OpenDCP saves the resulting j2k file somewhere else - it's now on a different physical drive, but still inside the same server.

    I now realise one of the potential problems is that I was using a disk in the storage pool which probably has an adverse affect on performance.

    In terms of random read / write, I suppose it is. I have, at most, 12 simultaneous accesses from remote machines (mapped drives) to read the files, compress them and write back the Jpeg2000 versions. I've now started writing the jp2k ones on a different physical drive (still inside the server), just in case it was a read/write issue.

    Since writing the original post, I've realised that there had been massive fragmentation as the result of writing the initial files. It's taken around 6 hours to half defragment! Having defragmented, it's slightly faster, but still slow, so I think it might be more an NTFS limitation than a physical fragmentation problem.

    I do have an EXT3 (maybe 4) Windows driver somewhere that works fine. How the performance is though, I don't know. I guess I'd be better with a small Linux NAS storage box to hold the data, as I can't afford to wipe my existing WHS installation. Thanks for the suggestion Peter

    Jim, I'd love to use SSDs, however, 2TB+ worth isn't going to be cheap
    If you using Adobe media encoder to take a video, crop, resize and output it as individual frames, then another piece of software convert those frames and combine them into one file then it looks to me like you're stuck deciding between using the glass bottle or the old shoe.
    http://weblogs.asp.net/alex_papadimo...25/408925.aspx

    I suspect there is software that does all you are asking of it with all of these steps internally.
    How are others acheiving what you are attempting?
    "In a perfect world... spammers would get caught, go to jail, and share a cell with many men who have enlarged their penises, taken Viagra and are looking for a new relationship."

Page 1 of 2 12 LastLast

Thread Information

Users Browsing this Thread

There are currently 1 users browsing this thread. (0 members and 1 guests)

Similar Threads

  1. Replies: 63
    Last Post: 14-11-2011, 09:17 AM
  2. cannot open applications. Registry problem?
    By roman.pro in forum Software
    Replies: 9
    Last Post: 21-06-2009, 05:30 PM
  3. best disk cloning software?
    By DsW in forum Software
    Replies: 30
    Last Post: 02-07-2008, 09:01 PM
  4. Nero vision express saying:'Burn process failed'
    By johnnr892 in forum Help! Quick Relief From Tech Headaches
    Replies: 15
    Last Post: 11-12-2005, 11:43 PM
  5. Script to compare number in text files?
    By wasabi in forum Software
    Replies: 5
    Last Post: 09-05-2005, 04:10 PM

Posting Permissions

  • You may not post new threads
  • You may not post replies
  • You may not post attachments
  • You may not edit your posts
  •