WhatsApp at (+91-9098855509) Support
ijprems Logo
  • Home
  • About Us
    • Editor Vision
    • Editorial Board
    • Privacy Policy
    • Terms & Conditions
    • Publication Ethics
    • Peer Review Process
  • For Authors
    • Publication Process(up)
    • Submit Paper Online
    • Pay Publication Fee
    • Track Paper
    • Copyright Form
    • Paper Format
    • Topics
  • Fees
  • Indexing
  • Conference
  • Contact
  • Archieves
    • Current Issue
    • Past Issue
  • More
    • FAQs
    • Join As Reviewer
  • Submit Paper

Recent Papers

Dedicated to advancing knowledge through rigorous research and scholarly publication

  1. Home
  2. Recent Papers

A Review of Neural Network Compression and Pruning using Shapley Pruning

Nandani Tiwari Tiwari

Download Paper

Paper Contents

Abstract

Neural network compression and pruning are two of the most vital methodologies that would make these deep architectures efficient so they can be deployed on these low-resource devices without significant deterioration in the chances of accuracy. Deep learning models are increasingly dominating NLP, computer vision, and healthcare applications and thus have witnessed an exponential increase in the related computational demands on memory for this model. This reduces the model sizes because either the precision of weight is downgraded or complexity in architecture is streamlined. In contrast, pruning results in the removal of excess weights, neurons, or even layer targeting the less impactful components a neural network. These optimizations resulted in better inference speed as well as lesser energy consumption which are crucial for this deployment in mobile platforms, Internet of things devices, and edge-computing systems. However, one of the challenges of this model simplification is the balance between efficiency and performance. Aggressive pruning and compression yield accuracy loss. Recent work on hardware-aware pruning, dynamic sparsity, and automated architecture design via NAS attempts to solve these challenges to produce more adaptive and high-performance models. Advancements in neural network compression and pruning are eyed to increase the scalability and sustainability of artificial intelligence technologies. These developments are expected to increase penetration across industries while helping the sector meet a growing need for energy efficiency in AI solutions.

Copyright

Copyright © 2024 Nandani Tiwari. This is an open access article distributed under the Creative Commons Attribution License.

Paper Details
Paper ID: IJPREMS41200008477
ISSN: 2321-9653
Publisher: ijprems
Page Navigation
  • Abstract
  • Copyright
About IJPREMS

The International Journal of Progressive Research in Engineering, Management and Science is a peer-reviewed, open access journal that publishes original research articles in engineering, management, and applied sciences.

Quick Links
  • Home
  • About Our Journal
  • Editorial Board
  • Publication Ethics
Contact Us
  • IJPREMS - International Journal of Progressive Research in Engineering Management and Science, motinagar, ujjain, Madhya Pradesh., india
  • Chat with us on WhatsApp: +91 909-885-5509
  • Email us: editor@ijprems.com
  • Sun-Sat: 9:00 AM - 9:00 PM

© 2025 International Journal of Progressive Research in Engineering, Management and Science. All Rights Reserved.

Terms & Conditions | Privacy Policy | Publication Ethics | Peer Review Process | Contact Us