- Community Home
- >
- Storage
- >
- HPE SimpliVity
- >
- how to clean up Omnistack /core 100% used.
Categories
Company
Local Language
Forums
Discussions
Forums
- Data Protection and Retention
- Entry Storage Systems
- Legacy
- Midrange and Enterprise Storage
- Storage Networking
- HPE Nimble Storage
Discussions
Discussions
Discussions
Forums
Discussions
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
- BladeSystem Infrastructure and Application Solutions
- Appliance Servers
- Alpha Servers
- BackOffice Products
- Internet Products
- HPE 9000 and HPE e3000 Servers
- Networking
- Netservers
- Secure OS Software for Linux
- Server Management (Insight Manager 7)
- Windows Server 2003
- Operating System - Tru64 Unix
- ProLiant Deployment and Provisioning
- Linux-Based Community / Regional
- Microsoft System Center Integration
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Community
Resources
Forums
Blogs
- Subscribe to RSS Feed
- Mark Topic as New
- Mark Topic as Read
- Float this Topic for Current User
- Bookmark
- Subscribe
- Printer Friendly Page
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
03-31-2019 08:00 PM
03-31-2019 08:00 PM
Hello,
Anyone here know how to clean up /core disk?
Filesystem Size Used Avail Use% Mounted on
udev 56G 4.0K 56G 1% /dev
tmpfs 12G 8.3M 12G 1% /run
/dev/sda1 9.1G 5.2G 3.5G 60% /roroot
tmpfs-root 56G 164M 56G 1% /tmproot
overlay 56G 164M 56G 1% /
/root/dev/dm-3 453M 2.3M 423M 1% /mnt
/root/dev/sda2 923M 92M 768M 11% /var/log
none 4.0K 0 4.0K 0% /sys/fs/cgroup
none 5.0M 0 5.0M 0% /run/lock
none 56G 1.1M 56G 1% /run/shm
none 100M 0 100M 0% /run/user
/dev/sdb1 30G 165M 29G 1% /cfgdb
/dev/sdb2 40G 819M 39G 3% /ctrdb
/dev/sdb3 4.8G 27M 4.8G 1% /pico
/dev/sda3 524M 6.4M 518M 2% /boot/efi
/dev/mapper/vg0-core 101G 101G 0 100% /core
Solved! Go to Solution.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
03-31-2019 08:12 PM
03-31-2019 08:12 PM
Re: how to clean up Omnistack /core 100% used.
That is not something it is safe to be playing with. If you are getting capacity warnings from the system you should open a support call.
While you CAN poke around inside the OVC now, that doesn't mean it is safe to do so. I have no idea what that LVM is used for, and the fact that it is full, absent an alarm, might not be an issue.
I work for HPE
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
03-31-2019 08:24 PM
03-31-2019 08:24 PM
Re: how to clean up Omnistack /core 100% used.
Thanks John for the reply.
New case 5337480575 has been logged for OVC - are running out of disk space and they advise below.
recommended to take a backup of the file through winscp, delete the folder which will give us disk space in OVC and recreate it .
I'm hesistant with the solution they provide as it may break OVC VM.
.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
04-01-2019 01:52 AM
04-01-2019 01:52 AM
Re: how to clean up Omnistack /core 100% used.
/core is used for dumps in the event of a crash the files gernerated are used for diagnostic purposes and will persist across reboots.
The advice is good these files may be deleted without issue.Not sure if i would bother winscping the files most of them are old and the other logs used for diagnosise will most likely have rotated,viewing the cores in isolation will not give a full picture.
I am an HPE employee
[Any personal opinions expressed are mine, and not official statements on behalf of Hewlett Packard Enterprise]
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
04-02-2019 07:19 PM