Next Article in Journal
An Interpolation-Based Evolutionary Algorithm for Bi-Objective Feature Selection in Classification
Previous Article in Journal
A Hybrid Intelligent Optimization Algorithm Based on a Learning Strategy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Intelligent Retrieval and Secure Acquisition of Embedded Data

1
Center for Cyber Security and Electronic Evidence, Fujian Police College, Fuzhou 350007, China
2
School of Mathematics, Shandong University, Jinan 250100, China
*
Author to whom correspondence should be addressed.
Mathematics 2024, 12(16), 2571; https://doi.org/10.3390/math12162571
Submission received: 29 May 2024 / Revised: 31 July 2024 / Accepted: 8 August 2024 / Published: 20 August 2024

Abstract

:
Data has attributes, and synchronous changes between data and attributes are the basic characteristics of a type of data: under the condition of supplementing (deleting) some attributes, some data elements are deleted (supplemented) from data. The attributes of data elements follow the conjunctive normal form. Based on the dynamic mathematical model P-sets, firstly, the conception and generation criteria of embedded data are proposed, and the inference theory of embedded data are studied. Secondly, intelligent retrieval and retrieval theorems for embedded data are proposed. Then, intelligent retrieval and secure acquisition algorithms for embedded data are designed. Finally, an application example is presented.
MSC:
03E72; 37N35; 93A10; 15B52

1. Introduction

Reference [1] adds a “dynamic characteristic” and proposes the concept of P-sets (packet sets) in the finite ordinary set (classical Cantor set), which makes up for the deficiency of the classical set only having a “static characteristic”; it creates P-sets, widely used in dynamic information problems such as data intelligent mining and obtaining, health big data, financial risk identification and so on. Several researchers have investigated the theory and application of p-sets and have achieved some outstanding results: The function P-sets is a functional form of the P-sets [2]. The concept of a P-augmented matrix based on P-sets is proposed and its application is given [3]. P-sets have an algebraic model [4]. A new method of intelligent information retrieval in dynamic information systems is present with the use of P-sets [5,6]. References [7,8,9] put forward the concept of P-augmented matrix based on P-sets, and gives the application of the P-augmented matrix in the dynamic findings of unknown information. More dynamic characteristics and applications of P-sets are discussed [10,11,12].
Formally, P-sets are the set pair ( X F ¯ , X F ) formed by the internal P-set X F ¯ and the outer P-set X F . In P-sets, the attribute α i of element x i satisfies the “conjunction normal form” in mathematical logic. Under certain conditions, p-sets can be reduced to ordinary sets of elements. Using P-sets, this paper presents the theoretical research of embedded data, and applies it to the problem of intelligent retrieval and secure acquisition of data.
In the study of unknown data acquisition and its application, there exists a type of data ( x ) which has attribute set α and dynamic characteristics. The dynamic characteristics of ( x ) are as follows:
  • If some attributes are supplemented within α , some data elements x i are deleted within ( x ) , and ( x ) generates internal embedded data ( x ) F ¯ , ( x ) F ¯ ( x ) ;
  • If some attributes are deleted within α , ( x ) is supplemented with some data elements x j , and ( x ) generates outer embedded data ( x ) F , ( x ) ( x ) F ;
  • Under the condition that 1 and 2 coexist, ( x ) generates embedded data ( ( x ) F ¯ , ( x ) F ) . The logical characteristics of attribute α i of data element x i here are the same as those of the P-sets.
If no attribute is supplemented in α , then the internal embedded data ( x ) F ¯ is hidden in ( x ) . If no attribute is deleted in α , then the outer embedded data ( x ) F is hidden outside ( x ) . In this case, ( x ) F ¯ , ( x ) F and ( ( x ) F ¯ , ( x ) F ) become unknown data.
This article uses the P-sets dynamic mathematical model to introduce dynamic characteristics into ordinary sets, replacing the “statics” of ordinary sets with “dynamism”, then the embedded data inference is discussed, and the intelligent retrieval and acquisition theorem of embedded data is given; based on the above theories and the encryption–decryption algorithm of the ellipse curve model, the intelligent retrieval and secure acquisition algorithm of embedded data are designed, and the application is given. The application examples come from venture capital identification.
The purpose of this study is to characterize the essential characteristics of complex data systems and extract important information from them, as well as to use the dynamic changes of complex data systems to achieve data security authentication, provided new theories and methods for data security transmission and authentication.

2. Dynamic Model with Attribute Conjunction

Cantor set X = { x 1 , x 2 , , x q } U is given, and α = { α 1 , α 2 , , α k } V is the attribute set of X ; X F ¯ is called an internal P-set generated by X ; moreover,
X F ¯ = X X
X is the F ¯ -element deleted set of X ,
X = { x i | x i X , f _ ( x i ) = u i ¯ X , f _ F _ }
If attribute set α F of X F ¯ meets
α F = α { α i | f ( β i ) = α i α , f F }
in (1), X F ¯ ,   X F ¯ = { x 1 , x 2 , , x p } , p < q ; p , q N + ; in (3), β i V , β i ¯ α , f F changes β i into f ( β i ) = α i α .
X F is called an outer P-set generated by X ; moreover,
X F = X X +
X + is the F -element supplemented set of X ,
X + = { u i | u i U , u i ¯ X , f ( u i ) = x i X , f F }
If the attribute set α F ¯ of X F meets,
α F ¯ = α { β i | f _ ( α i ) = β i ¯ α , f _ F _ _ }
in (4), X F = { x 1 , x 2 , , x r } , q < r , q , r N + ; in (6), α i α , f _ F _ _ , α F ¯ .
The set pair which is composed of internal packet set X F ¯ and outer packet set X F is called P-sets generated by X ; moreover,
X = ( X F ¯ , X F )
{ ( X i F ¯ , X j F ) | i I , j J }
(8) is the family of P-sets generated by X , and I and J are indicator sets.
Proposition 1.
If  F = F ¯ = , then  ( X F ¯ , X F )  and  X  meet
( X F ¯ , X F ) F = F ¯ = = X
Proposition 2.
If  F = F ¯ = , then  { ( X i F ¯ , X j F ) | i I , j J }  and  X  meet
{ ( X i F ¯ , X j F ) | i I , j J } F = F ¯ = = X
Remark 1.  
  • U , V  is the finite element domain and the finite attribute, respectively;
  • f F  and  f _ F _ _  are element (attribute) transfer;  F = { f 1 , f 2 f n }  and  F _ _ = { f 1 _ , f 2 _ , , f n _ }  are the family of element (attribute) transfer;
  • f s  characteristics are as follows f  changes  u i  into  f ( u i ) = x i X  and changes  β i  into  f ( β i ) = α i α  ;
  • f ¯ s  characteristics are as follows f ¯  changes  x i  into  f ¯ ( x i ) = u i ¯ X  and changes  α i  into  f ¯ ( α i ) = β i ¯ α ;  α i
  • The dynamic characteristics of the internal P-set are the same as those of down-counter T = T 1 ;
  • The dynamic characteristics of outer P-set are the same as those of accumulator  T = T + 1  . Such as,  X 1 F = X X 1 + , let  X = X 1 F ,  X 2 F = X X 2 + = ( X 1 F X 1 + ) X 2 + , , so on.
Conclusion 1.
When we add attributes continuously to  α , an internal P-set chain  X n F ¯ X n 1 F ¯ X 2 F ¯ X 1 F ¯  is generated by  X ;
Conclusion 2.
When we delete attributes continuously in  α , an outer P-set chain  X 1 F X 2 F X n 1 F X n F  is generated by  X ;
Conclusion 3.
When we add attributes and deleted attributes at the same time in  α , the P-sets chain  ( X n F ¯ , X 1 F ) ( X n 1 F ¯ , X 2 F ) ( X 2 F ¯ , X n 1 F ) ( X 1 F ¯ , X n F )  is dynamically generated by  X .
Figure 1 shows the two-dimensional distribution of the relationship between P-sets and the finite set of ordinary elements.
New concepts for the embedding are derived from Conclusions 1–3: X i F ¯ is the internal embedding generated by X , X i F ¯ X ; X j F is the outer embedding generated by X , X X j F ; ( X i F ¯ , X j F ) is the embedding generated by X , X i F ¯ X X j F .
Assumption 1.
( x ) = X , ( x ) F ¯ = X F ¯ ( x ) F = X F , ( ( x ) F ¯ , ( x ) F ) = ( X F ¯ , X F )   ( x ) , ( x ) F ¯ , ( x ) F  and  ( ( x ) F ¯ , ( x ) F )  are data,  x i ( x )  (or  x j ( x ) F ¯  or  x k ( x ) F ) is called a data element, and these symbols and names will be used directly without special explanation.

3. Embedded Generation of Data and Its Measure

( x ) F ¯ , ( x ) F are respectively referred to as the internally embedded data and externally embedded data generated by ( x ) , if ( x ) F ¯ , ( x ) F are satisfied, respectively,
car d ( ( x ) F ¯ ) car d ( ( x ) ) 0
car d ( ( x ) F ) car d ( ( x ) ) 0
The pair of embedded data formed by ( x ) F ¯ and ( x ) F is called embedded data ( x ) , which is denoted as:
( ( x ) F ¯ , ( x ) F )
{ ( ( x ) i F ¯ , ( x ) j F ) | i I , j J } is the embedded data family generated by ( x ) .
γ F ¯ , γ F are called the internal embedded measure and outer embedded measure of ( x ) F ¯ , ( x ) F , respectively, if γ F ¯ , γ F satisfies
γ F ¯ = car d ( ( x ) F ¯ ) / car d ( ( x ) )
γ F = car d ( ( x ) F ) / car d ( ( x ) )
( γ F ¯ , γ F ) is the embedded measure of ( ( x ) F ¯ , ( x ) F )
Where, in (11), (12) and (14), (15), c a r d = c a r d i n a l   n u m b e r .
Propositions 3–5 are obtained from Equations (11)–(15).
Proposition 3.
The necessary and sufficient condition for  ( x ) k F ¯  to be internal embedded data of  ( x )  is that the internal embedded measure  γ k F ¯  of  ( x ) k F ¯  is the inner point of the unit discrete interval (0,1], that is
γ k F ¯ ( 0 , 1 ]
Proposition 4.
The necessary and sufficient condition for  ( x ) k F ¯  to be outer embedded data of  ( x )  is that the outer embedded measure  γ k F ¯  of  ( x ) k F ¯  is the outer point of the unit discrete interval (0,1], that is
γ k F ¯ ( 0 , 1 ]
Proposition 5.
The necessary and sufficient condition for  ( ( x ) k F ¯ , ( x ) k F )  to be embedded data of  ( x )  is that the discrete interval  [ γ k F ¯ , γ k F ]  formed by the internal embedded measure  γ k F ¯  of  ( x ) k F ¯  and the outer embedded measure  γ k F ¯  of  ( x ) k F  and the unit discrete interval (0,1] satisfy:
[ γ k F ¯ , γ k F ] ( 0 , 1 ]
where (0,1] in Equations (16)–(18) are discrete intervals composed of numerical values 0 and  1 = γ = car d ( ( x ) ) / car d ( ( x ) ) , and  γ = 1  is the embedded measure of  ( x )  itself.
The proof of Proposition 4 and 5 is obtained directly from Equations (11)–(15), and the proof is omitted.
Theorem 1.
(Attribute theorem for internal embedded data) There exists an attribute set  Δ α  , such that the attribute set α F  of  ( x ) F ¯  and the attribute set  α  of  ( x )  satisfy:
α ( α F Δ α ) =
The proof of Theorem 1 is directly obtained from Equations (1)–(3); moreover, Δ α = { α i | f ( β i ) = α i α , f F } .
Theorem 2.
(Attribute theorem for outer embedded data) There exists an attribute set  α  , such that the attribute set α F ¯  of  ( x ) F  and the attribute set  α  of  ( x )  satisfy:
( α F ¯ α ) α =
The proof of Theorem 2 is directly obtained from Equations (4)–(6); moreover, α = { β i | f ¯ ( α i ) = β i ¯ α , f ¯ F ¯ } .
Inference 1.
There exists an attribute set  Δ α ,   α , such that the attribute set  ( α F , α F ¯ )  of  ( ( x ) F ¯ , ( x ) F )  and the attribute set  α  of  ( x )  satisfy:
( α F Δ α , α F ¯ α ) α =
where Formula (21) represents  ( α F Δ α ) α = , ( α F ¯ α ) α = .
The generation criteria for embedded data obtained from Propositions 3–5, Theorem 1–2, and Inference 1 are as follows.
(1)
Generation criteria for internal embedded data:
The implicit attribute α t α of data element x i ( x ) is the explicit attribute α t α k F of data element x i ( x ) k F ¯ .
Where ( x ) = { x 1 , x 2 , , x m } , ( x ) k F ¯ = { x 1 , x 2 , , x n } , n < m ; α = { α 1 , α 2 , , α t 1 } , α k F = { α 1 , α 2 , ,   α t 1 , α t } , α and α k F are the set of attributes for ( x ) and ( x ) k F ¯ , respectively.
(2)
Generation criteria for outer embedded data:
The implicit attribute α t α of data element x i ( x ) k F is the explicit attribute α t ¯ α k F ¯ of data element x i ( x ) k F .
Where ( x ) = { x 1 , x 2 , , x m } , ( x ) k F = { x 1 , x 2 , , x n } , m < n ; α = { α 1 , α 2 , , α t } , α k F ¯ = { α 1 , α 2 , , α t 1 } , α and α k F ¯ are the set of attributes for ( x ) and ( x ) k F , respectively.

4. Embedded Reasoning and Intelligent Retrieval of Embedded Data

If ( x ) , ( x ) k F ¯ and α , α k F satisfy
If α α k F , then ( x ) k F ¯ ( x ) .
Equation (22) is called internal embedded reasoning generated by embedded data, α α k F is called the internal embedded reasoning condition, ( x ) k F ¯ ( x ) is called the internal embedded reasoning conclusion. Where in (22), “ ” is equal to “ ”, α and α k F are the set of attributes for ( x ) and ( x ) k F ¯ , respectively.
If ( x ) , ( x ) k F and α , α k F ¯ satisfy
If α k F ¯ α , then ( x ) ( x ) k F .
Equation (23) is called outer embedded reasoning generated by embedded data, α k F ¯ α is called the outer embedded reasoning condition, ( x ) ( x ) k F is called the outer embedded reasoning conclusion; α and α k F are the set of attributes for ( x ) and ( x ) k F ¯ , respectively.
If ( ( x ) k F ¯ , ( x ) ) , ( ( x ) , ( x ) k F ) and ( α , α k F ¯ ) , ( α k F , α ) satisfy
If ( α , α k F ¯ ) ( α k F , α ) , then ( ( x ) k F ¯ , ( x ) ) ( ( x ) , ( x ) k F )
Equation (24) is referred to as embedded reasoning generated by embedded data, ( α , α k F ¯ ) ( α k F , α ) is referred to as the embedded reasoning condition, ( ( x ) k F ¯ , ( x ) ) ( ( x ) , ( x ) k F ) is referred to as the embedded reasoning conclusion.
Where Equation (24) indicates that both “ If α α k F , then ( x ) k F ¯ ( x ) . ” and “ If α k F ¯ α , then ( x ) ( x ) k F . ” are satisfied. The recursive form of Equation (24) is “ If ( α k F , α k + 1 F ¯ ) ( α k + 1 F , α k F ¯ ) , then ( ( x ) k + 1 F ¯ , ( x ) k F ) ( ( x ) k F ¯ , ( x ) k + 1 F ) ”, where ( α k F , α k + 1 F ¯ ) is the set of attributes of ( ( x ) k F ¯ , ( x ) k + 1 F ) .
From Equations (22)–(24), we can get Theorems 3–5 and Inferences 1–2.
Theorem 3.
If  ( x ) , ( x ) k F ¯  and  α , α k F  satisfy Equation (22), then
(1)
The internal embedded data ( x ) k F ¯  is intelligently retrieved and obtained within  ( x ) ;  ( x ) , ( x ) k F ¯  and  ( x )  satisfy:
( x ) k F ¯ = ( x ) ( x )
( x )  is redundant data composed of data elements  x i  within  ( x ) .
(2)
The attribute sets  α  of  ( x )  and  α k F  of  ( x ) k F ¯  satisfy:
α k F α
(3)
The implicit attribute α i  of data element  x i ( x ) k F ¯  is exposed.
Proof: 
(1) Using Equations (1)–(3) and (22), under the condition of α α k F , some elements x i in ( x ) = { x 1 , x 2 , , x t , x t + 1 , , x n } are deleted from ( x ) , which form ( x ) , and ( x ) generates ( x ) k F ¯ , moreover ( x ) k F ¯ , ( x ) and ( x ) satisfy ( x ) k F ¯ = ( x ) ( x ) , or ( x ) k F ¯ ( x ) , Equation (25) is obtained.
(2) Using Equations (1)–(3) in Section 2, it can be obtained that, assuming α = { α 1 , α 2 , , α λ } is the set of attributes of ( x ) , and x i ( x ) has the attribute α 1 , α 2 , , α λ , or in other words, α 1 , α 2 , , α λ is the explicit attribute of x i , then the attribute of x i satisfies the normal form of attribute conjunction α 1 α 2 α λ = t = 1 λ α t ; ( x ) and ( x ) have the same set of attributes: that is, α = α , and ( x ) k F ¯ has an attribute set α k F = { α 1 , α 2 , , α λ ,   α λ + 1 , , α n } , so it is obvious that α k F , α and α satisfy α k F α = α k F α , thus we get Equation (26).
(3) Under the condition of α α k F , the attribute α λ + 1 , α λ + 2 , , α n of x i ( x ) k F ¯ changes from implicit to explicit. □
Theorem 4.
If ( x ) , ( x ) k F  and  α , α k F ¯  satisfy Equation (23), then
(1)
The outer embedded data  ( x ) k F  is intelligently retrieved and obtained outside  ( x )  ;  ( x ) , ( x ) k F  and  Δ ( x )  satisfy:
( x ) k F = ( x ) Δ ( x )
Here,  Δ ( x )  is composed of missing data element  x i  within  ( x ) , which is a supplement within  ( x ) .
(2)
The attribute sets  Δ α  of  Δ ( x )  and  α k F ¯  of  ( x ) k F  satisfy:
α k F ¯ Δ α =
(3)
The explicit attribute of data element  x i ( x ) k F  is hidden.
The proof of Theorem 4 is similar to the proof of Theorem 3, which is omitted here.
Theorem 5.
If   ( ( x ) k F ¯ , ( x ) ) , ( ( x ) , ( x ) k F )  and  ( α , α k F ¯ ) , ( α k F , α )  satisfy Equation (24), then
(1)
The embedded data  ( ( x ) k F ¯ , ( x ) k F )  is intelligently retrieved, moreover
( ( x ) k F ¯ , ( x ) k F ) = ( ( x ) ( x ) , ( x ) Δ ( x ) )
(2)
The attribute set  α  of  ( x )  and the attribute set  Δ α  of  Δ ( x )  satisfy (26) and (28), respectively.
Inference 2.
The internal embedded data  ( x ) k F ¯  satisfying the recursive form internal embedded inference “ If α k F α k + 1 F , then ( x ) k + 1 F ¯ ( x ) k F ¯ . ” are intelligently retrieved in  ( x )  in order of  k = 1 , 2 , , n  , and they form the internal embedded data chain
( x ) n F ¯ ( x ) n 1 F ¯ ( x ) 2 F ¯ ( x ) 1 F ¯
Inference 3.
The outer embedded data  ( x ) k F  satisfying the recursive form outer embedded inference “ If α k + 1 F ¯ α k F ¯ , then ( x ) k F ( x ) k + 1 F . ” are intelligently retrieved outside  ( x )  in order of  k = 1 , 2 , , n  , and they form the outer embedded data chain
( x ) 1 F ( x ) 2 F ( x ) n 1 F ( x ) n F
Special note: Theorem 3–5 and Inference 2–3 are hidden in the application research of data retrieval, data filtering and dynamic data analysis. Equation (22)–(31) provide theoretical preparation for intelligent retrieval and discovery of unknown data.

5. Secure Acquisition Algorithm for Embedded Data

5.1. Ellipse Curve and Its Structure

The following equation is proposed by Koblitz N [13,14]:
y 2 = x 3 + a x + b
the E ( k ) determined by Equation (32) is called an ellipse curve as follows
E ( k ) = ( x , y ) x , y k , p ( x , y ) = 0 O
where discriminant D = ( 4 a 3 + 27 b 2 ) mod m 0 , k is a number field, a , b k ; O is the infinite point, and m is a prime number. Let be the additive operation of a point on E ( k ) . For any two points P ( x 1 , y 1 ) , Q ( x 1 , y 1 ) E ( k ) , there is a point R ( x 3 , y 3 ) on E ( k ) generated by P ( x 1 , y 1 ) Q ( x 1 , y 1 ) .
  • If P ( x 1 , y 1 ) Q ( x 2 , y 2 ) , then
    λ = y 2 y 1 x 2 x 1
    x 3 = ( λ 2 x 1 x 2 ) mod m
    y 3 = ( λ ( x 1 x 3 ) y 1 ) mod m
  • If P ( x 1 , y 1 ) = Q ( x 2 , y 2 ) , then
    λ = 3 x 1 2 + a 2 y 1
    x 3 = ( λ 2 x 1 x 2 ) mod m
    y 3 = ( λ ( x 1 x 3 ) y 1 ) mod m
Assumption 2.
The internal embedded data  ( x ) F ¯  in definition 4 is plaintext, denoted by  P m , namely  P m = ( x ) F ¯ ;  C m  is the ciphertext of  P m ;  A  is the encryptor of  P m  and the sender of  C m ;  B  is the decryptor of  C m  and the recipient of  P m . Both  A  and  B  jointly choose the same elliptical curve  E ( K ) , where  D 0 , and the same base point  G E ( k ) .

5.2. Encryption of Plaintext P m  and Decryption of Ciphertext C m

I. A chooses n A N + , and n A is the private key to A ; moreover,
P A = n A G
P A is the public key of A , then A gives the ciphertext C m of P m as follows:
C m = { k G , P m + k P B }
Then A sends C m to B .
II. B receives C m , and chooses n B N + ; n B is the private key to B ; moreover,
P B = n B G
P B is the public key of B ; B takes C m and calculates
P m + k P B n B ( k G ) = P m + k P B k ( n B G ) = P m + k P B k P B = P m
thus B obtains plaintext P m .
If B is the encryptor of P m and the sender of C m , and A is the receiver, decryptor and receiver of C m , then the process is similar to Equations (34)–(41).
Theorem 6.
(Abel Group Generation Theorem). If  E ( K )  is a set of points on an ellipse curve, and   is a point addition operation on  E ( K )  , then  < E ( K ) , >  is an Abel group.
Theorem 7.
(Closed loop theorem for generating base point  G ). If the base point  G  is taken on  E ( K ) , and   is the point addition operation on  E ( K ) , then there exists a positive integer  n , and the base point  G  generates a closed loop  D .
Theorem 8.
(Start-end theorem) If  D  is a closed loop generated by the base point  G  over  E ( K )  , then the base point  G  is taken at  D  ,  G G  ;  G  and the end point  G  satisfies
G = G
Theorem 6–8 is a direct result of ellipse curve; the proof is omitted.
From Equation (39), it can be seen that A obtains C m and sends it to B , while A must send it for signature to B . After B accepts A s signature, it provides authentication of the signature. At this point, B can make two choices: accept C m or reject C m .

5.3. Sending and Authentication of Signatures

Assumption 3.
A  is the provider of signature,  B  is the recipient and authenticator of the signature.  A ,  B  jointly choose elliptical curve  E ( K ) , D 0 , a prime number  n , and basic point  P E ( K ) . A  chooses the secret key  d ,  1 < d < n 1 , and computes  Q = d P . ( E , P , Q , n )  is public to  A ,  B . m is the signature plaintext chosen by  A ;  ( r , s )  is the signature ciphertext provided by  A .
  • A gives the signature ( r , s ) of m and sends it to B .
A chooses: 1 < k < n 1 ; k is the conversation key.
A calculates:
R = ( x R , y R ) = K P
r = x R mod n
S = k 1 ( h ( m ) + d r ) mod n
( r , s ) is the signature of m, ( r , s ) is sent to B .
Where h ( m ) is a hash function, r , s 0 .
2.
B accepts the signature ( r , s ) of m and makes an authentication;
A gives:
W = r 1 mod n
u 1 = h ( m ) w mod n
u 2 = r w mod n
B calculates:
R = ( x R , y R ) = u 1 P + u 2 Q
r = x R mod n
B makes the authentication: if r = r , then ( r , s ) is a legal signature sent by A , not from others; B accepts C m ; from (41), it can be seen that B obtains P m .
Theorem 9.
(Signature authentication theorem). Suppose  R  is any point on  E ( K ) ,  α , β N + ; if n  is a prime number, then there is a unique  R = α P + β Q , and it satisfies
x R mod n = x R mod n
here, R = ( x R , y R ) , R = ( x R , y R ) , and P , Q are two fixed points, P , Q E ( K ) .
Theorem 9 is obtained from Equations (43)–(50); its proof is omitted.

5.4. Security Acquisition Algorithm for Intelligent Retrieval of Embedded Data

This article only provides a secure acquisition algorithm for internal embedded data in intelligent retrieval, and its flowchart is shown in Figure 2. A secure retrieval algorithm for outer embedded data in intelligent retrieval is omitted.
Input: ( x ) , α , E ( K ) ; n is a large prime number and P E ( K ) is a basic point; output: internal embedded data ( x ) k F ¯ and secure authentication results obtained.
  • Based on ( x ) and α , generate data ( x ) F ¯ database and attribute α F database;
  • Generate recursive inference database with internal embedded data;
  • Generate internal embedded data ( x ) k F ¯ , k = 1 , 2 , n ;
  • Given the standard ( x ) F ¯ , , compare ( x ) k F ¯ with ( x ) F ¯ , ; if ( x ) k F ¯ ( x ) F ¯ , , then return 2; if ( x ) k F ¯ = ( x ) F ¯ , , then proceed to the next step;
  • Suppose plaintext P m = ( x ) k F ¯ ; A encrypts P m to obtain ciphertext C m , and sends C m and signature ( r , s ) to B ;
  • B certification signature ( r , s ) ; If r = r , then proceed to the next step, otherwise end, B cannot obtain P m ;
  • If r = r , then B accepts C m and obtains P m .
Algorithm structure and process
In Figure 2, the internal embedded data ( x ) F ¯ is subjected to recursive inference and compared with the standard internal embedded data ( x ) F ¯ , to obtain ( x ) k F ¯ that satisfies ( x ) k F ¯ = ( x ) F ¯ , . The internal embedded data ( x ) F ¯ completes the secure acquisition under the condition that the ellipse curve E ( K ) is satisfied. Among them, ( x ) k F ¯ is plaintext P m , that is P m = ( x ) k F ¯ , and C m is the ciphertext of P m . This algorithm is a combination of dynamic intelligent data retrieval algorithms and data security acquisition algorithms.

6. Application of Embedded Intelligent Data Retrieval and Secure Acquisition

For the sake of simplicity and generality, only the application of intelligent retrieval and secure acquisition of internal embedded data is given here. The example comes from the investment risk estimation in the financial system. The funds are invested in multiple companies, and the investor wants to know which companies will return satisfactory profits.

6.1. Given Basic Data and Attribute Sets and Selection of Investment Companies

Suppose that the basic data of the investment company x i selected by the investor is ( x ) , and α is the set of attributes that x i can obtain profits from. Moreover,
( x ) = { x 1 , x 2 , x 3 , x 4 , x 5 , x 6 , x 7 , x 8 }
α = { α 1 , α 2 , α 3 }
Here, α j α represents x i s ability to generate profits. The names of x i ( x ) and α i α are omitted.
At time t k T , if the investor acquires new attributes α 4 , α 5 , then α 4 , α 5 are embedded into α , and α generates α k F , moreover
α k F = α { α 4 , α 5 } = { α 1 , α 2 , α 3 , α 4 , α 5 }
Under the condition of α α k F , using the inference (22) of internal embedded data, we get: x 4 , x 6 , x 7 are deleted from ( x ) , then the unknown embedded data ( x ) k F ¯ is intelligently retrieved in ( x ) , and
( x ) k F ¯ = ( x ) { x 4 , x 6 , x 7 } = { x 1 , x 2 , x 3 , x 5 , x 8 }
At time t k + λ T , if the investor acquires new attributes α 6 , then α generates α k F , and
α k + λ F = α k F { α 6 } = { α 1 , α 2 , α 3 , α 4 , α 5 , α 6 }
Under the condition of α k F α k + λ F , ( x ) k + λ F ¯ is obtained by the recursive inference form “ If α k F α k + λ F , then ( x ) k + λ F ¯ ( x ) k F ¯ , and
( x ) k + λ F ¯ = ( x ) k F ¯ { x 2 } = { x 1 , x 3 , x 5 , x 8 }
Here, Equation (57) indicates that under the condition of α k F α k + λ F , x 2 is deleted from ( x ) k F ¯ and ( x ) k + λ F ¯ is intelligently obtained within ( x ) k F ¯ .
In Equations (52)–(57), the attribute α i of x i ( x ) , the attribute α j of x j ( x ) k F ¯ , and the attribute α k c of x k ( x ) k + λ F ¯ satisfy the normal form of attribute conjunction α i = t = 1 3 α t , α j = t = 1 3 α t α 4 α 5 = t = 1 5 α t , and α k = t = 1 5 α t α 6 = t = 1 6 α t , respectively.
The ( x ) k + λ F ¯ in (57) represents the profitable company determined by the investor, and the estimated profits obtained by x 1 , x 3 , x 5 , x 8 are shown in Table 1.
In Table 1, y i ( y ) k + λ F ¯ represents the profit value of x i ( x ) k + λ F ¯ .

6.2. Secure Acquisition and Confirmation of True Profit Value

Assume A is the investor determined by investment company D , and B is the profit determiner determined by investment company D . In Table 1, ( y ) k + λ F ¯ = { y 1 , y 3 , y 5 , y 8 } = { ( 9 , 7 ) , ( 13 , 16 ) } is defined as plaintext P m = { P m , 1 , P m , 2 } = { ( 9 , 7 ) , ( 13 , 16 ) } ; C m = { C m , 1 , C m , 2 } is the ciphertext of P m ; y 2 = x 3 + a x + b mod m = x 3 + x + 1 mod 23 is an elliptical curve jointly determined by A and B . According to Equations (34)–(37) in Section 4, we get:
E ( K ) = { ( 0 , 1 ) , ( 0 , 22 ) , ( 1 , 7 ) , ( 1 , 16 ) , ( 3 , 10 ) , ( 3 , 13 ) , ( 4 , 0 ) , ( 5 , 4 ) , ( 5 , 19 ) , ( 6 , 4 ) , ( 6 , 19 ) , ( 7 , 11 ) , ( 7 , 12 ) , ( 9 , 7 ) , ( 9 , 16 ) , ( 11 , 3 ) , ( 11 , 20 ) , ( 12 , 4 ) , ( 12 , 19 ) , ( 13 , 17 ) , ( 13 , 16 ) , ( 17 , 3 ) , ( 17 , 20 ) , ( 18 , 3 ) , ( 18 , 20 ) , ( 19 , 5 ) , ( 19 , 18 ) } ;
27 G + G = O ; G = ( 5 , 4 ) E ( K ) is the base point chosen by A and B ; n A = 5 , n B = 17 , P A = n A G = ( 7 , 11 ) ,   P B = n B G = ( 18 , 3 ) .
Using Equation (39), A takes plaintext P m , 1 = ( 9 , 7 ) , k = 3 , P B = ( 18 , 3 ) and obtains ciphertext C m , 1 ; and, using Equations (43)–(45), A provides the signature ( r , s ) :
C m , 1 = { k G , P m , 1 + k P B } = { 3 G , 14 G + 72 G } = { 3 G , 86 G } = { 3 G , 2 G } = { ( 18 , 3 ) , ( 0 , 22 ) }
A sends C m , 1 and ( r , s ) to B .
After receiving the ciphertext C m , 1 and signature ( r , s ) , B determines r = r , if so, then the authentication signature ( r , s ) comes from A ; B decrypts C m , 1 using (40) and obtains:
P m , 1 + k P B n B ( k G ) = 2 G 17 ( 24 G ) = 2 G 408 G = 2 G 16 G = 14 G = 14 G = ( 9 , 7 ) = P m , 1
Thus, B obtains the true profit values of x 1 , x 3 as y 1 , y 3 .The encryption–decryption process of P m , 2 is similar to P m , 1 , and is omitted.
Special note: Equation (57) and ( y ) k + λ F ¯ are trade secrets and should be confirmed by the investment company manager. A uses the intelligent retrieval algorithm in Figure 2 to obtain Equation (57) and ( y ) k + λ F ¯ , and then sends them to the manager using the secure retrieval algorithm in Figure 2. The manager provides authentication for Equation (57) and ( y ) k + λ F ¯ through the algorithm.

7. Conclusions

In application research fields such as data retrieval, data recognition, data hiding, and data camouflage, data ( x ) always exists simultaneously with its attribute set α (the feature set of data elements). The dynamic changes in data (decreases or increases in data elements within ( x ) ) are synchronized with the dynamic changes of attributes (increases or decreases of attributes within α ). Dynamic data often appears in the research of artificial intelligence systems, while static data is only a special case for dynamic data. References [15,16] only provide application research on static data using classical methods. Considering the dynamic characteristics of data in practical applications, this paper presents an application study of a class of dynamic data using the dynamic mathematical model P-sets as the research tool. This article proposes the concept of embedded data, provides a reasoning theory and intelligent retrieval of embedded data, designs intelligent retrieval and secure acquisition algorithms for embedded data, and provides their applications.
This paper is just the beginning of the research on dynamic data problems. As the research on the application of inverse P-sets, function P-sets and function inverse P-sets in data retrieval, data analysis, data image segmentation-hiding, data image camouflage-reduction and other problems, the research is still in progress. Inverse P-sets [17,18] are obtained by improving P-sets, which are the dual model of P-sets. In inverse P-sets, the attributes α i of data element x i meet the attribute disjunctive characteristics. Function inverse P-sets [19,20] are obtained by replacing the internal data element x ¯ i of (x) with a function s ¯ ( x ) i and improving inverse P-sets. The logical characteristics of function inverse P-sets are the same as those of inverse P-sets. These dynamic mathematical models provide theoretical and methodological support for intelligent data retrieval, discovery of unknown data, and intelligent mining of big data applications.

Author Contributions

Conceptualization, S.C.; methodology, S.C.; writing—original draft preparation, S.C.; writing—review and editing, S.C.; software, S.C.; investigation and supervision, K.S.; validation and funding acquisition, K.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was sponsored by the Fujian Provincial Department of Science and Technology Project under Grant 2020H6024.

Data Availability Statement

The datasets used and analysed during the current study are available from the corresponding author on reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Shi, K. P-sets. J. Shandong Univ. Nat. Sci. 2008, 43, 77–84. (In Chinese) [Google Scholar]
  2. Shi, K. Function P-sets. J. Shandong Univ. Nat. Sci. 2011, 46, 62–69. (In Chinese) [Google Scholar] [CrossRef]
  3. Zhang, X. P-augmented matrix and its application in dynamic tracking recognition. J. Anhui Univ. Nat. Sci. 2022, 46, 53–58. (In Chinese) [Google Scholar]
  4. Li, X. An algebraic model of P-sets. J. Shangqiu Norm. Univ. 2020, 36, 1–5. (In Chinese) [Google Scholar]
  5. Zhang, X.; Shen, L. Intelligent separation and identification of sub-information based on dynamic mathematical model. Symmetry 2023, 15, 477. [Google Scholar] [CrossRef]
  6. Zhang, X.; Li, X. P-information fusion and its P-matrix reasoning intelligent generation. J. Shandong Univ. Nat. Sci. 2017, 52, 93–99. (In Chinese) [Google Scholar]
  7. Fan, C.; Huang, S. Inverse P-reasoning discovery ienification of inverse P-information. Intercont. J. Digit. Content Technol. Appl. 2012, 6, 735–744. [Google Scholar]
  8. Zhang, X.; Shen, L. P-Matrix reasoning and information intelligent mining. Appl. Math. Nonlinear Sci. 2023, 8, 29–38. [Google Scholar] [CrossRef]
  9. Zhang, L.; Ren, X. The relationship between abnormal information system and inverse P-augmented matrices. J. Shandong Univ. Nat. Sci. 2019, 54, 15–21. (In Chinese) [Google Scholar]
  10. Lin, H.; Fan, C. The dual form of P-reasoning and identification of unknown attribute. Int. J. Digit. Content Technol. Appl. 2012, 6, 121–131. [Google Scholar]
  11. Hao, X.; Li, N. Quantitative characteristics and applications of P-information hidden mining. J. Shandong Univ. Nat. Sci. 2019, 54, 9–14. (In Chinese) [Google Scholar]
  12. Fan, C.; Lin, H. P-sets and the reasoning-identification of disaster information. Int. J. Converg. Inf. Technol. 2012, 7, 337–345. [Google Scholar]
  13. Koblitz, N. Elliptic curve cryptosystems. Math. Comput. 1987, 48, 203–209. [Google Scholar] [CrossRef]
  14. Menezes, A.J.; Vanstone, S.A. Elliptic curve cryptosystems and their implementation. Crypotology 1993, 6, 209–224. [Google Scholar] [CrossRef]
  15. Xie, F.; Xu, Y.; Ma, M.; Pang, X. A safe acceleration method for multi-task twin support vector machine. Int. J. Mach. Learn. Cybern. 2022, 13, 1713–1728. [Google Scholar] [CrossRef]
  16. Yang, T.; Wang, J.; Sun, Q.; Hibar, D.P.; Jahanshad, N.; Liu, L.; Wang, Y.; Zhan, L.; Thompson, P.M.; Ye, J. Detection genetic risk factors for Alzheimer’s disease in whole genome sequence data via lasso Screening. In Proceedings of the IEEE 2015 12th International Symposium on Biomedical Imaging, Brooklyn, NY, USA, 16–19 April 2015; IEEE: New York, NY, USA, 2015; pp. 985–989. [Google Scholar]
  17. Yu, X.; Xu, F. Function P (σ,τ)-set and its characteristics. J. Jilin Univ. Nat. Sci. 2018, 56, 53–59. (In Chinese) [Google Scholar]
  18. Zhang, Y. Random function inverse P-sets and its characteristics depending on attributes. J. Shandong Univ. Nat. Sci. 2014, 49, 90–94. (In Chinese) [Google Scholar]
  19. Shi, K. Function inverse P-sets and the hiding information generated by function inverse P-information law fusion. In Proceedings of the Conference on e-Business, e-Services and e-Society I3E 2014: Digital Services and Information Intelligence, Sanya, China, 28–30 November 2014; Springer: Berlin/Heidelberg, Germany, 2014; pp. 224–237. [Google Scholar]
  20. Zhang, N.; Ren, X. The intelligent data mining-classification and its dynamic management. J. Minnan Norm. Univ. Nat. Sci. 2020, 33, 103–107. (In Chinese) [Google Scholar]
Figure 1. X F ¯ , X F are represented by solid lines, and finite ordinary element set X is represented by dotted lines.
Figure 1. X F ¯ , X F are represented by solid lines, and finite ordinary element set X is represented by dotted lines.
Mathematics 12 02571 g001
Figure 2. Flowchart of internal embedded data intelligent retrieval and secure acquisition algorithm.
Figure 2. Flowchart of internal embedded data intelligent retrieval and secure acquisition algorithm.
Mathematics 12 02571 g002
Table 1. Estimates of the value of x 1 , x 3 , x 5 , x 8 acquired profits.
Table 1. Estimates of the value of x 1 , x 3 , x 5 , x 8 acquired profits.
( x ) k + λ F ¯ x 1 x 3 x 5 x 8
( y ) k + λ F ¯ 971316
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, S.; Shi, K. Intelligent Retrieval and Secure Acquisition of Embedded Data. Mathematics 2024, 12, 2571. https://doi.org/10.3390/math12162571

AMA Style

Chen S, Shi K. Intelligent Retrieval and Secure Acquisition of Embedded Data. Mathematics. 2024; 12(16):2571. https://doi.org/10.3390/math12162571

Chicago/Turabian Style

Chen, Shuzhen, and Kaiquan Shi. 2024. "Intelligent Retrieval and Secure Acquisition of Embedded Data" Mathematics 12, no. 16: 2571. https://doi.org/10.3390/math12162571

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop