尿道炎症吃什么药好| 其多列是什么意思| 部分导联t波改变是什么意思| 肩膀疼吃什么药| 代表什么意思| 35是什么意思| 吞咽困难挂什么科| 毛囊炎长什么样| 光年是什么单位| 西洋参和花旗参有什么区别| 中筛是检查什么项目| 天天晚上睡觉做梦是什么原因| 频繁做梦是什么原因| 陈皮的功效与作用主要治什么病| 儿加一笔是什么字| 试金石什么意思| 补肾最好的药是什么药| 皮肤粗糙缺什么维生素| 四月十四日是什么节日| 口臭是什么原因导致的呢| 拿什么证明分居两年| 摩尔每升是什么单位| 麦露迪女鞋什么档次| 什么样的头发| 夏天喝什么解暑| 蛋白粉什么时候喝| 天高地厚是什么生肖| 西药是用什么材料做的| 一天中什么时候最热| 瑞五行属什么| 壬寅年五行属什么| 总胆固醇低是什么原因| 孙红雷的老婆叫什么名字| 白细胞低什么原因| 流涎是什么意思| 庚午日是什么意思| 萎缩性阴道炎用什么药| 脸上长红色的痘痘是什么原因| 正月十九是什么日子| rosa是什么意思| 腱子肉是什么意思| 血脂血糖高吃什么食物好| 山东简称为什么是鲁不是齐| 94年属什么今年多大| 颈椎退变是什么意思| 芬必得是什么药| 寻麻疹涂抹什么药膏| 什么时候敷面膜是最佳时间| 什么样的男人不能嫁| 心悸心慌吃什么药| 吃益生菌有什么好处| 左眼一直跳有什么预兆| 葡萄球菌感染是什么原因引起的| 鸡宝是什么| 92什么意思| 藏在我回忆里的那个人什么歌| 军训是什么时候开始的| 孕酮低对胎儿有什么影响| 鼠肚鸡肠是什么生肖| 血糖高什么原因引起| 能量守恒是什么意思| 茉莉花长什么样| 百忧解是什么药| 舌头发白什么原因| 9月24号什么星座| 有个马的标志是什么车| 月桂酰两性基乙酸钠是什么| 胃寒可以吃什么水果| 白带多是什么原因| 5年生存率是什么意思| 鼓的偏旁部首是什么| 孢子阳性是什么意思| 调岗是什么意思| 亲子是什么意思| 销魂是什么意思| 去心火喝什么茶好| 逸搏心律什么意思| 早上八点多是什么时辰| 公开遴选公务员是什么意思| 农历五月初五是什么节日| 皮肤变白是什么原因| 月经来了有血块是什么原因| 达瓦里氏什么意思| 男性生殖器官叫什么| 慢性胃炎用什么药效果最好| 最机灵的动物是什么生肖| 白陶土样便见于什么病| 做生意的人最忌讳什么| 五月初是什么星座| 爸爸的姐姐叫什么| 罴是什么动物| 没晨勃说明什么问题| 手指缝脱皮是什么原因| 又什么又什么的什么| 头皮痒用什么洗头好| 公鸡为什么打鸣| 肌肉拉伤吃什么药| 把脉把的是什么脉| 子宫脱垂什么症状| 急性子是什么意思| ccu病房是什么意思| 八一年属什么生肖| 猫为什么不怕蛇| 咳嗽有白痰吃什么药| 琥珀是什么意思| 印堂发亮预兆着什么| 小康生活的标准是什么| 北京为什么叫四九城| 绊倒是什么意思| 蓝莓树长什么样| hbcag是什么意思| 番薯是什么| 吃布洛芬有什么副作用| 盼头是什么意思| 习惯是什么意思| 补气血喝什么口服液好| 醉代表什么生肖| 剧透是什么意思| 怀孕什么时候能测出来| gi值是什么| 牛蛙和青蛙有什么区别| 激素六项是查什么的| 尿常规能检查出什么| 病毒性肝炎有什么症状| 绿壳鸡蛋是什么鸡生的| 针灸后需要注意什么| 孕妇脚抽筋是什么原因| 胃酸胃胀反酸水吃什么药| 三个毛念什么字| 蚕除了吃桑叶还能吃什么| 葫芦五行属什么| 吃稀饭配什么菜好吃| 医师是什么级别| 半熟芝士是什么意思| 子宫内膜增厚是什么原因| 抗心磷脂抗体是什么意思| 姜字五行属什么| 开门是什么意思| 吃杏子有什么好处| 生姜什么时候种| 什么是假性狐臭| 关系是什么意思| 脾湿吃什么中成药| 外阴灼热用什么药| 笑气是什么气体| 失眠吃什么药| 卵巢检查做什么项目| 外贸是什么| 炖牛肉放什么容易烂| 一岁宝宝流鼻涕吃什么药| 前列腺增大是什么原因| 520是什么节日| 脚后跟干裂用什么药膏| 吃什么下火效果最好| 什么的鱼| 儿童中耳炎用什么药最好| 霉菌性中耳炎用什么药| 九月初九是什么节日| 女人能日到什么时候| 金银花主治什么| 黄豆芽炒什么好吃| 欲什么意思| 氯中毒吃什么可以解毒| 辛辣指的是什么| ac什么意思| 口头禅是什么意思| 梦见搬家是什么预兆| bf是什么| 特异性生长因子指什么| 绿色食品是什么意思| 澳门什么时候回归祖国| 凌晨是什么时辰| 小腿肌肉痛什么原因| 小孩子手脱皮是什么原因引起的| 失眠为什么开奥氮平片| 肚子痛拉肚子吃什么药| 吃胡萝卜有什么好处| 日主是什么意思| 切是什么偏旁| 猫爪草长什么样| 防风通圣颗粒治什么病| 蜂蜜不能和什么一起吃| 舌头发涩是什么原因造成的| 前位子宫是什么意思| kp是什么| 市政府办公室主任是什么级别| 丙磺舒是什么药| 卡其色是什么颜色| 营养师属于什么专业| 清炖鸡汤放什么调料| 宫颈肥大有什么危害| 搬家当天有什么讲究| 提拉米苏是什么东西| 凤毛麟角是什么意思| 新疆有什么特产| 巳是什么意思| 肝结节是什么病严重吗| 嘴子是什么意思| 什么东西清肺最好| 午餐肉是什么肉| 一什么叮咛| 氯气是什么颜色| 脑门长痘痘是什么原因| 什么原因引起低压高| 肠癌是什么症状| 一天两包烟会导致什么后果| 偶数和奇数是什么意思| 梦呓是什么意思| 血塞通治什么病| 醛固酮高吃什么降压药| 晚上睡觉手麻是什么原因| 舌头下面的筋叫什么| 与其让你在我怀中枯萎是什么歌| 万力什么字| 跳蚤咬了擦什么药最好| 额头长闭口是什么原因| 什么的饰品| 禾花鱼是什么鱼| 15天来一次月经是什么原因| n是什么牌子的鞋| 孱弱是什么意思| 耳朵痒用什么药| e6是什么意思| 玩手机头疼是什么原因| 好雅兴是什么意思| 送女朋友什么礼物| 梦见抓蛇是什么预兆| 丙肝是什么病严重吗| 儿童身份证需要什么材料| 紫苏什么味道| 滑胎是什么意思| 传染病八项挂什么科| 什么是阳光抑郁症| 雪纳瑞什么颜色最贵| 肿瘤病人不能吃什么| 不打狂犬疫苗会有什么后果| 口干口臭什么原因引起的| 棕色用什么颜色调出来| 断掌是什么意思| 厚颜无耻是什么生肖| 舌根白苔厚是什么原因| 涤纶是什么布料| 寸脉弱是什么原因| 过敏有什么症状| 尿酸高中医叫什么病| 什么是单亲家庭| 黄金芽是什么茶| 一个火一个旦读什么字| 下午一点到三点是什么时辰| 什么又什么| 什么床品牌最好| 什么牌子的益生菌调理肠胃比较好| 附睾炎吃什么药| 紧急避孕药对身体有什么伤害| leu是什么氨基酸| 嘴角烂了是什么原因| 日单是什么意思| 胡椒粉是什么做的| 月亮是什么| 柳下惠姓什么| 水晶消磁是什么意思| 白羊座的幸运色是什么颜色| 梦见狗咬自己是什么意思| 脊髓炎是什么病| 百度

北京发布空气重污染“橙警”

百度 努力构建一流城市学理论研究中心、学术交流中心、信息发布中心、人才培养中心和协同创新中心的定位,积极打造集评选、论坛、平台、课题、人才、宣传、基金、基地、培训、咨询等“十位一体”的城市学研究链,建立健全“模块化研究、矩阵式管理”的运行模式,打好打响“杭州牌”、“浙江牌”、“中国牌”和“国际牌”,构建“党政、企事业、智库、媒体、市民”五位一体的城市学研究复合主体,在探索中国特色新型智库建设上形成自身特色、凸显优势。

Digital image processing is the use of a digital computer to process digital images through an algorithm.[1][2] As a subcategory or field of digital signal processing, digital image processing has many advantages over analog image processing. It allows a much wider range of algorithms to be applied to the input data and can avoid problems such as the build-up of noise and distortion during processing. Since images are defined over two dimensions (perhaps more), digital image processing may be modeled in the form of multidimensional systems. The generation and development of digital image processing are mainly affected by three factors: first, the development of computers;[3] second, the development of mathematics (especially the creation and improvement of discrete mathematics theory);[4] and third, the demand for a wide range of applications in environment, agriculture, military, industry and medical science has increased.[5]

History

edit

Many of the techniques of digital image processing, or digital picture processing as it often was called, were developed in the 1960s, at Bell Laboratories, the Jet Propulsion Laboratory, Massachusetts Institute of Technology, University of Maryland, and a few other research facilities, with application to satellite imagery, wire-photo standards conversion, medical imaging, videophone, character recognition, and photograph enhancement.[6] The purpose of early image processing was to improve the quality of the image. It was aimed for human beings to improve the visual effect of people. In image processing, the input is a low-quality image, and the output is an image with improved quality. Common image processing include image enhancement, restoration, encoding, and compression. The first successful application was the American Jet Propulsion Laboratory (JPL). They used image processing techniques such as geometric correction, gradation transformation, noise removal, etc. on the thousands of lunar photos sent back by the Space Detector Ranger 7 in 1964, taking into account the position of the Sun and the environment of the Moon. The impact of the successful mapping of the Moon's surface map by the computer has been a success. Later, more complex image processing was performed on the nearly 100,000 photos sent back by the spacecraft, so that the topographic map, color map and panoramic mosaic of the Moon were obtained, which achieved extraordinary results and laid a solid foundation for human landing on the Moon.[7]

The cost of processing was fairly high, however, with the computing equipment of that era. That changed in the 1970s, when digital image processing proliferated as cheaper computers and dedicated hardware became available. This led to images being processed in real-time, for some dedicated problems such as television standards conversion. As general-purpose computers became faster, they started to take over the role of dedicated hardware for all but the most specialized and computer-intensive operations. With the fast computers and signal processors available in the 2000s, digital image processing has become the most common form of image processing, and is generally used because it is not only the most versatile method, but also the cheapest.

Image sensors

edit

The basis for modern image sensors is metal–oxide–semiconductor (MOS) technology,[8] invented at Bell Labs between 1955 and 1960,[9][10][11][12][13][14] This led to the development of digital semiconductor image sensors, including the charge-coupled device (CCD) and later the CMOS sensor.[8]

The charge-coupled device was invented by Willard S. Boyle and George E. Smith at Bell Labs in 1969.[15] While researching MOS technology, they realized that an electric charge was the analogy of the magnetic bubble and that it could be stored on a tiny MOS capacitor. As it was fairly straightforward to fabricate a series of MOS capacitors in a row, they connected a suitable voltage to them so that the charge could be stepped along from one to the next.[8] The CCD is a semiconductor circuit that was later used in the first digital video cameras for television broadcasting.[16]

The NMOS active-pixel sensor (APS) was invented by Olympus in Japan during the mid-1980s. This was enabled by advances in MOS semiconductor device fabrication, with MOSFET scaling reaching smaller micron and then sub-micron levels.[17][18] The NMOS APS was fabricated by Tsutomu Nakamura's team at Olympus in 1985.[19] The CMOS active-pixel sensor (CMOS sensor) was later developed by Eric Fossum's team at the NASA Jet Propulsion Laboratory in 1993.[20] By 2007, sales of CMOS sensors had surpassed CCD sensors.[21]

MOS image sensors are widely used in optical mouse technology. The first optical mouse, invented by Richard F. Lyon at Xerox in 1980, used a 5 μm NMOS integrated circuit sensor chip.[22][23] Since the first commercial optical mouse, the IntelliMouse introduced in 1999, most optical mouse devices use CMOS sensors.[24][25]

Image compression

edit

An important development in digital image compression technology was the discrete cosine transform (DCT), a lossy compression technique first proposed by Nasir Ahmed in 1972.[26] DCT compression became the basis for JPEG, which was introduced by the Joint Photographic Experts Group in 1992.[27] JPEG compresses images down to much smaller file sizes, and has become the most widely used image file format on the Internet.[28] Its highly efficient DCT compression algorithm was largely responsible for the wide proliferation of digital images and digital photos,[29] with several billion JPEG images produced every day as of 2015.[30]

Medical imaging techniques produce very large amounts of data, especially from CT, MRI and PET modalities. As a result, storage and communications of electronic image data are prohibitive without the use of compression.[31][32] JPEG 2000 image compression is used by the DICOM standard for storage and transmission of medical images. The cost and feasibility of accessing large image data sets over low or various bandwidths are further addressed by use of another DICOM standard, called JPIP, to enable efficient streaming of the JPEG 2000 compressed image data.[33]

Digital signal processor (DSP)

edit

Electronic signal processing was revolutionized by the wide adoption of MOS technology in the 1970s.[34] MOS integrated circuit technology was the basis for the first single-chip microprocessors and microcontrollers in the early 1970s,[35] and then the first single-chip digital signal processor (DSP) chips in the late 1970s.[36][37] DSP chips have since been widely used in digital image processing.[36]

The discrete cosine transform (DCT) image compression algorithm has been widely implemented in DSP chips, with many companies developing DSP chips based on DCT technology. DCTs are widely used for encoding, decoding, video coding, audio coding, multiplexing, control signals, signaling, analog-to-digital conversion, formatting luminance and color differences, and color formats such as YUV444 and YUV411. DCTs are also used for encoding operations such as motion estimation, motion compensation, inter-frame prediction, quantization, perceptual weighting, entropy encoding, variable encoding, and motion vectors, and decoding operations such as the inverse operation between different color formats (YIQ, YUV and RGB) for display purposes. DCTs are also commonly used for high-definition television (HDTV) encoder/decoder chips.[38]

Tasks

edit

Digital image processing allows the use of much more complex algorithms, and hence, can offer both more sophisticated performance at simple tasks, and the implementation of methods which would be impossible by analogue means.

In particular, digital image processing is a concrete application of, and a practical technology based on:

Some techniques which are used in digital image processing include:

Digital image transformations

edit

Filtering

edit

Digital filters are used to blur and sharpen digital images. Filtering can be performed by:

  • convolution with specifically designed kernels (filter array) in the spatial domain[39]
  • masking specific frequency regions in the frequency (Fourier) domain

The following examples show both methods:[40]

Filter type Kernel or mask Example
Original Image    
Spatial Lowpass    
Spatial Highpass    
Fourier Representation Pseudo-code:

image = checkerboard

F = Fourier Transform of image

Show Image: log(1+Absolute Value(F))

 
Fourier Lowpass    
Fourier Highpass    

Image padding in Fourier domain filtering

edit

Images are typically padded before being transformed to the Fourier space, the highpass filtered images below illustrate the consequences of different padding techniques:

Zero padded Repeated edge padded
   

Notice that the highpass filter shows extra edges when zero padded compared to the repeated edge padding.

Filtering code examples

edit

MATLAB example for spatial domain highpass filtering.

img=checkerboard(20);                           % generate checkerboard
% **************************  SPATIAL DOMAIN  ***************************
klaplace=[0 -1 0; -1 5 -1;  0 -1 0];             % Laplacian filter kernel
X=conv2(img,klaplace);                          % convolve test img with
                                                % 3x3 Laplacian kernel
figure()
imshow(X,[])                                    % show Laplacian filtered
title('Laplacian Edge Detection')

Affine transformations

edit

Affine transformations enable basic image transformations including scale, rotate, translate, mirror and shear as is shown in the following examples:[40]

Transformation Name Affine Matrix Example
Identity    
Reflection    
Scale    
Rotate     where θ = ?π/6? =30°
Shear    

To apply the affine matrix to an image, the image is converted to matrix in which each entry corresponds to the pixel intensity at that location. Then each pixel's location can be represented as a vector indicating the coordinates of that pixel in the image, [x, y], where x and y are the row and column of a pixel in the image matrix. This allows the coordinate to be multiplied by an affine-transformation matrix, which gives the position that the pixel value will be copied to in the output image.

However, to allow transformations that require translation transformations, 3-dimensional homogeneous coordinates are needed. The third dimension is usually set to a non-zero constant, usually 1, so that the new coordinate is [x, y, 1]. This allows the coordinate vector to be multiplied by a 3×3 matrix, enabling translation shifts. Thus, the third dimension, i.e. the constant 1, allows translation.

Because matrix multiplication is associative, multiple affine transformations can be combined into a single affine transformation by multiplying the matrix of each individual transformation in the order that the transformations are done. This results in a single matrix that, when applied to a point vector, gives the same result as all the individual transformations performed on the vector [x, y, 1] in sequence. Thus a sequence of affine transformation matrices can be reduced to a single affine transformation matrix.

For example, 2-dimensional coordinates only permit rotation about the origin (0, 0). But 3-dimensional homogeneous coordinates can be used to first translate any point to (0, 0), then perform the rotation, and lastly translate the origin (0, 0) back to the original point (the opposite of the first translation). These three affine transformations can be combined into a single matrix—thus allowing rotation around any point in the image.[41]

Image denoising with mathematical morphology

edit

Mathematical morphology (MM) is a nonlinear image processing framework that analyzes shapes within images by probing local pixel neighborhoods using a small, predefined function called a structuring element. In the context of grayscale images, MM is especially useful for denoising through dilation and erosion—primitive operators that can be combined to build more complex filters.

Suppose we have:

  • A discrete grayscale image:  
  • A structuring element:  

Here,   defines the neighborhood of relative coordinates   over which local operations are computed. The values of   bias the image during dilation and erosion.

Dilation
Grayscale dilation is defined as:

 

For example, the dilation at position (1, 1) is calculated as:

 

Erosion
Grayscale erosion is defined as:

 

For example, the erosion at position (1, 1) is calculated as:

 

Results

edit

After applying dilation to  :  

After applying erosion to  :  

Opening and Closing

edit

MM operations, such as opening and closing, are composite processes that utilize both dilation and erosion to modify the structure of an image. These operations are particularly useful for tasks such as noise removal, shape smoothing, and object separation.

  • Opening: This operation is performed by applying erosion to an image first, followed by dilation. The purpose of opening is to remove small objects or noise from the foreground while preserving the overall structure of larger objects. It is especially effective in situations where noise appears as isolated bright pixels or small, disconnected features.

For example, applying opening to an image   with a structuring element   would first reduce small details (through erosion) and then restore the main shapes (through dilation). This ensures that unwanted noise is removed without significantly altering the size or shape of larger objects.

  • Closing: This operation is performed by applying dilation first, followed by erosion. Closing is typically used to fill small holes or gaps within objects and to connect broken parts of the foreground. It works by initially expanding the boundaries of objects (through dilation) and then refining the boundaries (through erosion).

For instance, applying closing to the same image   would fill in small gaps within objects, such as connecting breaks in thin lines or closing small holes, while ensuring that the surrounding areas are not significantly affected.

Both opening and closing can be visualized as ways of refining the structure of an image: opening simplifies and removes small, unnecessary details, while closing consolidates and connects objects to form more cohesive structures.

Structuring element Mask Code Example
Original Image None Use Matlab to read Original image
original = imread('scene.jpg');
image = rgb2gray(original);
[r, c, channel] = size(image);
se = logical([1 1 1 ; 1 1 1 ; 1 1 1]);
[p, q] = size(se);
halfH = floor(p/2);
halfW = floor(q/2);
time = 3;           % denoising 3 times with all method
 
Original lotus
Dilation   Use Matlab to dilation
imwrite(image, "scene_dil.jpg")
extractmax = zeros(size(image), class(image));
for i = 1 : time
    dil_image = imread('scene_dil.jpg');
    for col = (halfW + 1): (c - halfW)
        for row = (halfH + 1) : (r - halfH)
            dpointD = row - halfH;
            dpointU = row + halfH;
            dpointL = col - halfW;
            dpointR = col + halfW;
            dneighbor = dil_image(dpointD:dpointU, dpointL:dpointR);
            filter = dneighbor(se);
            extractmax(row, col) = max(filter);
        end
    end
    imwrite(extractmax, "scene_dil.jpg");
end
 
Denoising picture with dilation method
Erosion   Use Matlab to erosion
imwrite(image, 'scene_ero.jpg');
extractmin = zeros(size(image), class(image));
for i = 1: time
    ero_image = imread('scene_ero.jpg');
    for col = (halfW + 1): (c - halfW)
        for row = (halfH +1): (r -halfH)
            pointDown = row-halfH;
            pointUp = row+halfH;
            pointLeft = col-halfW;
            pointRight = col+halfW;
            neighbor = ero_image(pointDown:pointUp,pointLeft:pointRight);
            filter = neighbor(se);
            extractmin(row, col) = min(filter);
        end
    end
    imwrite(extractmin, "scene_ero.jpg");
end
 
Opening   Use Matlab to Opening
imwrite(extractmin, "scene_opening.jpg")
extractopen = zeros(size(image), class(image));
for i = 1 : time
    dil_image = imread('scene_opening.jpg');
    for col = (halfW + 1): (c - halfW)
        for row = (halfH + 1) : (r - halfH)
            dpointD = row - halfH;
            dpointU = row + halfH;
            dpointL = col - halfW;
            dpointR = col + halfW;
            dneighbor = dil_image(dpointD:dpointU, dpointL:dpointR);
            filter = dneighbor(se);
            extractopen(row, col) = max(filter);
        end
    end
    imwrite(extractopen, "scene_opening.jpg");
end
 
Closing   Use Matlab to Closing
imwrite(extractmax, "scene_closing.jpg")
extractclose = zeros(size(image), class(image));
for i = 1 : time
    ero_image = imread('scene_closing.jpg');
    for col = (halfW + 1): (c - halfW)
        for row = (halfH + 1) : (r - halfH)
            dpointD = row - halfH;
            dpointU = row + halfH;
            dpointL = col - halfW;
            dpointR = col + halfW;
            dneighbor = ero_image(dpointD:dpointU, dpointL:dpointR);
            filter = dneighbor(se);
            extractclose(row, col) = min(filter);
        end
    end
    imwrite(extractclose, "scene_closing.jpg");
end
 
Denoising picture with closing method

Applications

edit

Digital camera images

edit

Digital cameras generally include specialized digital image processing hardware – either dedicated chips or added circuitry on other chips – to convert the raw data from their image sensor into a color-corrected image in a standard image file format. Additional post processing techniques increase edge sharpness or color saturation to create more naturally looking images.

Film

edit

Westworld (1973) was the first feature film to use the digital image processing to pixellate photography to simulate an android's point of view.[42] Image processing is also vastly used to produce the chroma key effect that replaces the background of actors with natural or artistic scenery.

Face detection

edit
 
Face detection process

Face detection can be implemented with mathematical morphology, the discrete cosine transform (DCT), and horizontal projection.

General method with feature-based method

The feature-based method of face detection is using skin tone, edge detection, face shape, and feature of a face (like eyes, mouth, etc.) to achieve face detection. The skin tone, face shape, and all the unique elements that only the human face have can be described as features.

Process explanation

  1. Given a batch of face images, first, extract the skin tone range by sampling face images. The skin tone range is just a skin filter.
    1. Structural similarity index measure (SSIM) can be applied to compare images in terms of extracting the skin tone.
    2. Normally, HSV or RGB color spaces are suitable for the skin filter. E.g. HSV mode, the skin tone range is [0,48,50] ~ [20,255,255]
  2. After filtering images with skin tone, to get the face edge, morphology and DCT are used to remove noise and fill up missing skin areas.
    1. Opening method or closing method can be used to achieve filling up missing skin.
    2. DCT is to avoid the object with skin-like tone. Since human faces always have higher texture.
    3. Sobel operator or other operators can be applied to detect face edge.
  3. To position human features like eyes, using the projection and find the peak of the histogram of projection help to get the detail feature like mouth, hair, and lip.
    1. Projection is just projecting the image to see the high frequency which is usually the feature position.

Improvement of image quality method

edit

Image quality can be influenced by camera vibration, over-exposure, gray level distribution too centralized, and noise, etc. For example, noise problem can be solved by smoothing method while gray level distribution problem can be improved by histogram equalization.

Smoothing method

In drawing, if there is some dissatisfied color, taking some color around dissatisfied color and averaging them. This is an easy way to think of Smoothing method.

Smoothing method can be implemented with mask and convolution. Take the small image and mask for instance as below.

image is  

mask is  

After convolution and smoothing, image is  

Observing image[1, 1], image[1, 2], image[2, 1], and image[2, 2].

The original image pixel is 1, 4, 28, 30. After smoothing mask, the pixel becomes 9, 10, 9, 9 respectively.

new image[1, 1] =   * (image[0,0]+image[0,1]+image[0,2]+image[1,0]+image[1,1]+image[1,2]+image[2,0]+image[2,1]+image[2,2])

new image[1, 1] = floor(  * (2+5+6+3+1+4+1+28+30)) = 9

new image[1, 2] = floor({  * (5+6+5+1+4+6+28+30+2)) = 10

new image[2, 1] = floor(  * (3+1+4+1+28+30+7+3+2)) = 9

new image[2, 2] = floor(  * (1+4+6+28+30+2+3+2+2)) = 9

Gray Level Histogram method

Generally, given a gray level histogram from an image as below. Changing the histogram to uniform distribution from an image is usually what we called histogram equalization.

 
Figure 1
 
Figure 2

In discrete time, the area of gray level histogram is  (see figure 1) while the area of uniform distribution is  (see figure 2). It is clear that the area will not change, so  .

From the uniform distribution, the probability of   is   while the  

In continuous time, the equation is  .

Moreover, based on the definition of a function, the Gray level histogram method is like finding a function   that satisfies f(p)=q.

Improvement method Issue Before improvement Process After improvement
Smoothing method noise

with Matlab, salt & pepper with 0.01 parameter is added
to the original image in order to create a noisy image.

 
  1. read image and convert image into grayscale
  2. convolution the grayscale image with the mask  
  3. denoisy image will be the result of step 2.
 
Histogram Equalization Gray level distribution too centralized
 
Refer to the Histogram equalization
 

Challenges

edit
  1. Noise and Distortions: Imperfections in images due to poor lighting, limited sensors, and file compression can result in unclear images that impact accurate image conversion.
  2. Variability in Image Quality: Variations in image quality and resolution, including blurry images and incomplete details, can hinder uniform processing across a database.
  3. Object Detection and Recognition: Identifying and recognising objects within images, especially in complex scenarios with multiple objects and occlusions, poses a significant challenge.
  4. Data Annotation and Labelling: Labelling diverse and multiple images for machine recognition is crucial for further processing accuracy, as incorrect identification can lead to unrealistic results.
  5. Computational Resource Intensity: Accessing adequate computational resources for image processing can be challenging and costly, hindering progress without sufficient resources.

See also

edit

References

edit
  1. ^ Chakravorty, Pragnan (2018). "What is a Signal? [Lecture Notes]". IEEE Signal Processing Magazine. 35 (5): 175–177. Bibcode:2018ISPM...35e.175C. doi:10.1109/MSP.2018.2832195. S2CID 52164353.
  2. ^ Gonzalez, Rafael (2018). Digital image processing. New York, NY: Pearson. ISBN 978-0-13-335672-4. OCLC 966609831.
  3. ^ Nagornov, Nikolay N.; Lyakhov, Pavel A.; Bergerman, Maxim V.; Kalita, Diana I. (2024). "Modern Trends in Improving the Technical Characteristics of Devices and Systems for Digital Image Processing". IEEE Access. 12: 44659–44681. Bibcode:2024IEEEA..1244659N. doi:10.1109/ACCESS.2024.3381493. ISSN 2169-3536.
  4. ^ Yamni, Mohamed; Daoui, Achraf; Abd El-Latif, Ahmed A. (February 2024). "Efficient color image steganography based on new adapted chaotic dynamical system with discrete orthogonal moment transforms". Mathematics and Computers in Simulation. 225: 1170–1198. doi:10.1016/j.matcom.2024.01.023.
  5. ^ Hung, Che-Lun (28 May 2020). "Computational Algorithms on Medical Image Processing". Current Medical Imaging. 16 (5): 467–468. doi:10.2174/157340561605200410144743. PMID 32484080.
  6. ^ Azriel Rosenfeld, Picture Processing by Computer, New York: Academic Press, 1969
  7. ^ Gonzalez, Rafael C. (2008). Digital image processing. Woods, Richard E. (Richard Eugene), 1954– (3rd ed.). Upper Saddle River, N.J.: Prentice Hall. pp. 23–28. ISBN 978-0-13-168728-8. OCLC 137312858.
  8. ^ a b c Williams, J. B. (2017). The Electronics Revolution: Inventing the Future. Springer. pp. 245–8. ISBN 978-3-319-49088-5.
  9. ^ US2802760A, Lincoln, Derick & Frosch, Carl J., "Oxidation of semiconductive surfaces for controlled diffusion", issued 13 August 1957 
  10. ^ Frosch, C. J.; Derick, L (1957). "Surface Protection and Selective Masking during Diffusion in Silicon". Journal of the Electrochemical Society. 104 (9): 547. doi:10.1149/1.2428650.
  11. ^ KAHNG, D. (1961). "Silicon-Silicon Dioxide Surface Device". Technical Memorandum of Bell Laboratories: 583–596. doi:10.1142/9789814503464_0076. ISBN 978-981-02-0209-5. {{cite journal}}: ISBN / Date incompatibility (help)
  12. ^ Lojek, Bo (2007). History of Semiconductor Engineering. Berlin, Heidelberg: Springer-Verlag Berlin Heidelberg. p. 321. ISBN 978-3-540-34258-8.
  13. ^ Ligenza, J.R.; Spitzer, W.G. (1960). "The mechanisms for silicon oxidation in steam and oxygen". Journal of Physics and Chemistry of Solids. 14: 131–136. Bibcode:1960JPCS...14..131L. doi:10.1016/0022-3697(60)90219-5.
  14. ^ Lojek, Bo (2007). History of Semiconductor Engineering. Springer Science & Business Media. p. 120. ISBN 9783540342588.
  15. ^ James R. Janesick (2001). Scientific charge-coupled devices. SPIE Press. pp. 3–4. ISBN 978-0-8194-3698-6.
  16. ^ Boyle, William S; Smith, George E. (1970). "Charge Coupled Semiconductor Devices". Bell Syst. Tech. J. 49 (4): 587–593. Bibcode:1970BSTJ...49..587B. doi:10.1002/j.1538-7305.1970.tb01790.x.
  17. ^ Fossum, Eric R. (12 July 1993). "Active pixel sensors: Are CCDS dinosaurs?". In Blouke, Morley M. (ed.). Charge-Coupled Devices and Solid State Optical Sensors III. Proceedings of the SPIE. Vol. 1900. pp. 2–14. Bibcode:1993SPIE.1900....2F. CiteSeerX 10.1.1.408.6558. doi:10.1117/12.148585. S2CID 10556755.
  18. ^ Fossum, Eric R. (2007). "Active Pixel Sensors" (PDF). Eric Fossum. S2CID 18831792. Archived (PDF) from the original on 29 August 2019.
  19. ^ Matsumoto, Kazuya; et al. (1985). "A new MOS phototransistor operating in a non-destructive readout mode". Japanese Journal of Applied Physics. 24 (5A): L323. Bibcode:1985JaJAP..24L.323M. doi:10.1143/JJAP.24.L323. S2CID 108450116.
  20. ^ Fossum, Eric R.; Hondongwa, D. B. (2014). "A Review of the Pinned Photodiode for CCD and CMOS Image Sensors". IEEE Journal of the Electron Devices Society. 2 (3): 33–43. doi:10.1109/JEDS.2014.2306412.
  21. ^ "CMOS Image Sensor Sales Stay on Record-Breaking Pace". IC Insights. 8 May 2018. Archived from the original on 21 June 2019. Retrieved 6 October 2019.
  22. ^ Lyon, Richard F. (2014). "The Optical Mouse: Early Biomimetic Embedded Vision". Advances in Embedded Computer Vision. Springer. pp. 3–22 (3). ISBN 9783319093871.
  23. ^ Lyon, Richard F. (August 1981). "The Optical Mouse, and an Architectural Methodology for Smart Digital Sensors" (PDF). In H. T. Kung; Robert F. Sproull; Guy L. Steele (eds.). VLSI Systems and Computations. Computer Science Press. pp. 1–19. doi:10.1007/978-3-642-68402-9_1. ISBN 978-3-642-68404-3. S2CID 60722329. Archived (PDF) from the original on 26 February 2014.
  24. ^ Brain, Marshall; Carmack, Carmen (24 April 2000). "How Computer Mice Work". HowStuffWorks. Retrieved 9 October 2019.
  25. ^ Benchoff, Brian (17 April 2016). "Building the First Digital Camera". Hackaday. Retrieved 30 April 2016. the Cyclops was the first digital camera
  26. ^ Ahmed, Nasir (January 1991). "How I Came Up With the Discrete Cosine Transform". Digital Signal Processing. 1 (1): 4–5. Bibcode:1991DSP.....1....4A. doi:10.1016/1051-2004(91)90086-Z. Archived from the original on 10 June 2016. Retrieved 10 October 2019.
  27. ^ "T.81 – Digital compression and coding of continuous-tone still images – requirements and guidelines" (PDF). CCITT. September 1992. Archived (PDF) from the original on 17 July 2019. Retrieved 12 July 2019.
  28. ^ Svetlik, Joe (31 May 2018). "The JPEG image format explained". BT Group. Archived from the original on 5 August 2019. Retrieved 5 August 2019.
  29. ^ Caplan, Paul (24 September 2013). "What Is a JPEG? The Invisible Object You See Every Day". The Atlantic. Archived from the original on 9 October 2019. Retrieved 13 September 2019.
  30. ^ Baraniuk, Chris (15 October 2015). "JPeg lockdown: Restriction options sought by committee". BBC News. Archived from the original on 9 October 2019. Retrieved 13 September 2019.
  31. ^ Nagornov, Nikolay N.; Lyakhov, Pavel A.; Valueva, Maria V.; Bergerman, Maxim V. (2022). "RNS-Based FPGA Accelerators for High-Quality 3D Medical Image Wavelet Processing Using Scaled Filter Coefficients". IEEE Access. 10: 19215–19231. Bibcode:2022IEEEA..1019215N. doi:10.1109/ACCESS.2022.3151361. ISSN 2169-3536. S2CID 246895876. Medical imaging systems produce increasingly accurate images with improved quality using higher spatial resolutions and color bit-depth. Such improvements increase the amount of information that needs to be stored, processed, and transmitted.
  32. ^ Dhouib, D.; Na?t-Ali, A.; Olivier, C.; Naceur, M.S. (June 2021). "ROI-Based Compression Strategy of 3D MRI Brain Datasets for Wireless Communications". IRBM. 42 (3): 146–153. doi:10.1016/j.irbm.2020.05.001. S2CID 219437400. Because of the large amount of medical imaging data, the transmission process becomes complicated in telemedicine applications. Thus, in order to adapt the data bit streams to the constraints related to the limitation of the bandwidths a reduction of the size of the data by compression of the images is essential.
  33. ^ Xin, Gangtao; Fan, Pingyi (11 June 2021). "A lossless compression method for multi-component medical images based on big data mining". Scientific Reports. 11 (1): 12372. doi:10.1038/s41598-021-91920-x. ISSN 2045-2322. PMC 8196061. PMID 34117350.
  34. ^ Grant, Duncan Andrew; Gowar, John (1989). Power MOSFETS: theory and applications. Wiley. p. 1. ISBN 978-0-471-82867-9. The metal–oxide–semiconductor field-effect transistor (MOSFET) is the most commonly used active device in the very large-scale integration of digital integrated circuits (VLSI). During the 1970s these components revolutionized electronic signal processing, control systems and computers.
  35. ^ Shirriff, Ken (30 August 2016). "The Surprising Story of the First Microprocessors". IEEE Spectrum. 53 (9). Institute of Electrical and Electronics Engineers: 48–54. doi:10.1109/MSPEC.2016.7551353. S2CID 32003640. Archived from the original on 13 October 2019. Retrieved 13 October 2019.
  36. ^ a b "1979: Single Chip Digital Signal Processor Introduced". The Silicon Engine. Computer History Museum. Archived from the original on 3 October 2019. Retrieved 14 October 2019.
  37. ^ Taranovich, Steve (27 August 2012). "30 years of DSP: From a child's toy to 4G and beyond". EDN. Archived from the original on 14 October 2019. Retrieved 14 October 2019.
  38. ^ Stankovi?, Radomir S.; Astola, Jaakko T. (2012). "Reminiscences of the Early Work in DCT: Interview with K.R. Rao" (PDF). Reprints from the Early Days of Information Sciences. 60. Archived (PDF) from the original on 13 October 2019. Retrieved 13 October 2019.
  39. ^ Zhang, M. Z.; Livingston, A. R.; Asari, V. K. (2008). "A High Performance Architecture for Implementation of 2-D Convolution with Quadrant Symmetric Kernels". International Journal of Computers and Applications. 30 (4): 298–308. doi:10.1080/1206212x.2008.11441909. S2CID 57289814.
  40. ^ a b Gonzalez, Rafael (2008). Digital Image Processing, 3rd. Pearson Hall. ISBN 978-0-13-168728-8.
  41. ^ House, Keyser (6 December 2016). Affine Transformations (PDF). Foundations of Physically Based Modeling & Animation. A K Peters/CRC Press. ISBN 978-1-4822-3460-2. Archived (PDF) from the original on 30 August 2017. Retrieved 26 March 2019. {{cite book}}: |website= ignored (help)
  42. ^ A Brief, Early History of Computer Graphics in Film Archived 17 July 2012 at the Wayback Machine, Larry Yaeger, 16 August 2002 (last update), retrieved 24 March 2010

Further reading

edit
edit
广西属于什么方向 什么歌最好听 六亲不认是什么生肖 ifound是什么牌子 黄芪不适合什么人吃
全品类是什么意思 阑尾炎挂号挂什么科 什么网名好 百步穿杨是什么生肖 属蛇男和什么属相最配
梅花代表什么象征意义 水为什么是透明的 卷饼里面配什么菜好吃 直男癌是什么意思 秫米是什么米
芈月和嬴政什么关系 小蜗牛吃什么 女属羊和什么属相最配 早上5点多是什么时辰 gs什么意思
疖肿挂什么科hcv7jop6ns9r.cn 我们为什么会笑hcv8jop8ns2r.cn 16开是什么意思hcv7jop9ns6r.cn 精华液是干什么用的hcv9jop7ns0r.cn 雪莲菌泡牛奶有什么功效hcv8jop5ns8r.cn
橘色五行属什么hanqikai.com 游离三碘甲状腺原氨酸是什么意思hcv8jop9ns8r.cn 内分泌失调吃什么药效果最好hcv7jop7ns0r.cn 追随是什么意思jasonfriends.com 马陆吃什么hcv7jop9ns5r.cn
肝炎吃什么药好hcv9jop6ns5r.cn 游乐场都有什么项目hcv9jop8ns3r.cn 梦见自己吐血是什么征兆hcv7jop6ns3r.cn 乞丐是什么生肖hcv8jop5ns6r.cn 栀子花什么时候开hcv8jop5ns8r.cn
银杏果什么时候成熟hcv8jop7ns7r.cn 三个羊念什么hcv8jop7ns2r.cn 抗美援朝是什么时候hcv8jop4ns4r.cn 湿气重是什么意思hcv8jop8ns8r.cn 榻榻米是什么hcv8jop5ns9r.cn
百度