file count limit in a single directory

Discussion created by Patrick_Higgins_6715 on Oct 1, 2015
Latest reply on Oct 6, 2015 by Tran-Hoang_Hau_348
Been getting some intermittent performance problems and so have been looking for anything odd. One server and agent have had more issues, so started looking around. Running Oracle Linux 5.10, Kernel 2.6.18-371. 64bit File system is using ext3 with a 1024 block size, Found that one directory has over 100k files ...../agents/db_oracle/temp.   In a prior life, I had severe issues with FAT filesystems and over 10k files in a single directory - taking 5 minutes for a directory listing, check for file existence, or adding a new file.  Never had heard of similar issues with Linux. Found opposing views on the web and even from two different linux admins here.  - One viewpoint is that over 60k files in a single directory is bad for ext3 (ideal would be less than 10k).  Even if file system can handle it, shell might not. - Different view sees that over 90% of the inodes are available (2.7m used out of 67m), so file count is not relevant Any problems with UC4/Automic (AE and agents at v10.0.2)?